New hybrid conjugate gradient algorithms for unconstrained optimization

Size: px
Start display at page:

Download "New hybrid conjugate gradient algorithms for unconstrained optimization"

Transcription

1 ew hybrid conjugate gradient algorithms for unconstrained optimization eculai Andrei Research Institute for Informatics, Center for Advanced Modeling and Optimization, 8-0, Averescu Avenue, Bucharest, Romania, Abstract ew hybrid conjugate gradient algorithms are proposed and analyzed In these hybrid algorithms the famous parameter is computed as a convex combination of the Pola-Ribière-Polya and Dai-Yuan conjugate gradient algorithms In one hybrid algorithm the parameter in convex combination is computed in such a way that the conjugacy condition is satisfied, independent of the line search In the other, the parameter in convex combination is computed in such a way that the conjugate gradient direction is the ewton direction he algorithm uses the standard Wolfe line search conditions umerical comparisons with conjugate gradient algorithms using a set of 750 unconstrained optimization problems, some of them from the CUE library, show that the hybrid computational scheme based on conjugacy condition outperform the nown hybrid conjugate gradient algorithms MSC: 49M07, 49M0, 90C06, 65K Keywords: Unconstrained optimization, hybrid conjugate gradient method, conjugacy condition, numerical comparisons Introduction Let us consider the nonlinear unconstrained optimization problem n min f( x): x R, () { } n where f : R R is a continuously differentiable function, bounded from below For solving this problem, starting from an initial guess x0 generates a sequence { x } as R n, a nonlinear conjugate gradient method, x = x + + αd, () where α > 0 is obtained by line search, and the directions d are generated as d+ = g+ + s, d0 = g0 (3) In (3) is nown as the conjugate gradient parameter, s = x+ x and g = f( x) Consider the Euclidean norm and define y = g+ g he line search in the conjugate gradient algorithms often is based on the standard Wolfe conditions: f ( x + α d ) f ( x ) ρα g d (4), g+ d σ gd, (5) where d is a descent direction and 0< ρ σ < Plenty of conjugate gradient methods are nown, and an excellent survey of these methods, with a special attention on their global convergence, is given by Hager and Zhang [9] Different conjugate gradient algorithms correspond to different choices for the scalar parameter Some of these methods as

2 Fletcher and Reeves (FR) [6], Dai and Yuan (DY) [] and Conjugate Descent (CD) proposed by Fletcher [5]: FR g+ g+ DY g+ g+ CD g+ g+ =, =, =, g g gs have strong convergence properties, but they may have modest practical performance due to jamming On the other hand, the methods of Pola Ribière [3] and Polya (PRP) [4], Hestenes and Stiefel (HS) [0] or Liu and Storey (LS) []: PRP g+ y HS g+ y LS g+ y =,, =, = g g gs in general may not be convergent, but they often have better computational performances In this paper we focus on hybrid conjugate gradient methods hese methods are combinations of different conjugate gradient algorithms, mainly they being proposed to avoid the jamming phenomenon One of the first hybrid conjugate gradient algorithms has been introduced by ouati-ahmed and Storey [8], where the parameter is computed as: PRP g+ y PRP FR =, if 0, S g = FR g + =, otherwise g he PRP method has a built-in restart feature that directly addresses to jamming Indeed, when the step s is small, then the factor y in the numerator of tends to zero herefore, descent direction PRP becomes small and the search direction g + PRP d + is very close to the steepest Hence, when the iterations jam, the method of ouati-ahmed and Storey uses the PRP computational scheme Another hybrid conjugate gradient method was given by Hu and Storey [], where is: { { }} = max 0, min, HuS PRP FR in (3) As above, when the method of Hu and Storey is jamming, then the PRP method is used instead he combination between LS and CD conjugate gradient methods leads to the following hybrid method: { { }} = max 0, min, LS CD LS CD he CD method of Fletcher [5] is very close to FR method With an exact line search, CD method is identical to FR Similarly, for an exact line search, LS method is also identical to PRP herefore, the hybrid LS-CD method with an exact line search has similar performances with the hybrid method of Hu and Storey Gilbert and ocedal [7] suggested a combination between PRP and FR methods as: Since FR { { }} G max FR, min PRP, FR = is alwa nonnegative, it follows that G can be negative he method of Gilbert and ocedal has the same advantage of avoiding jamming Using the standard Wolfe line search, the DY method alwa generates descent directions and if the gradient is Lipschitz continuously the method is global convergent In an effort to improve their algorithm, Dai and Yuan [3] combined their algorithm with other conjugate gradient algorithm, proposing the following two hybrid methods: { c { }} hdy max DY, min HS, DY =,

3 { { }} = max 0, min,, hdyz HS DY where c = ( σ ) /( + σ ) For the standard Wolfe conditions (4) and (5), under the Lipschitz continuity of the gradient, Dai and Yuan [3] established the global convergence of these hybrid computational schemes In this paper we propose another hybrid conjugate gradient as a convex combination of PRP and DY conjugate gradient algorithms We selected these two methods to combine in a hybrid conjugate gradient algorithm because PRP has good computational properties, on one side, and DY has strong convergence properties, on the other side Often PRP method performs better in practice than DY and we speculate this in order to have a good practical conjugate algorithm he structure of the paper is as follows In section we introduce our hybrid conjugate gradient algorithm and prove that it generates descent directions satisfying in some conditions the sufficient descent condition Section 3 presents the algorithms and in section 4 we show its convergence analis In section 5 some numerical experiments and performance profiles of Dolan-Moré [4] corresponding to this new hybrid conjugate gradient algorithm and some other conjugate gradient algorithms are presented he performance profiles corresponding to a set of 750 unconstrained optimization problems in the CUE test problem library [6], as well as some other unconstrained optimization problems presented in [] show that this hybrid conjugate gradient algorithm outperform the nown hybrid conjugate gradient algorithms ew hybrid conjugate gradient algorithms he iterates x, x, x, the stepsize directions where d 0 of our algorithm are computed by means of the recurrence () where α > 0 is determined according to the Wolfe conditions (4) and (5), and the are generated by the rule: d = g + s, d0 = g0, (6) + + = ( θ ) + θ PRP DY g y g g = ( θ ) + g g θ (7) and θ is a scalar parameter satisfying 0 θ, which follows to be determined Observe PRP DY that if θ = 0, then =, and if θ =, then = On the other hand, if PRP DY 0< θ <, then is a convex combination of and Referring to the PRP method, Pola and Ribière [3] proved that when function f is strongly convex and the line search is exact, then the PRP method is global convergent In an effort to understand the behavior of the PRP method, Powell [5] showed that if the step length s = x x approaches to zero, the line search is exact and the gradient f ( x) is + Lipschitz continuous, then the PRP method is globally convergent Additionally, assuming that the search direction is a descent direction, Yuan [9] established the global convergence of the PRP method for strongly convex functions and a Wolfe line search For general nonlinear functions the convergence of the PRP method is uncertain Powell [6] gave a 3 dimensional example, in which the function to be minimized is not strongly convex, showing that even with an exact line search, the PRP method may not converge to a stationary point Later on Dai [7] presented another example this time with a strongly convex function for which the PRP method fails to generate a descent direction herefore, theoretically the convergence of the PRP method is limited to strongly convex functions For general nonlinear functions the convergence of the PRP method is established under restrictive conditions (Lipschitz continuity, exact line search and the stepsize tends to zero) However, the numerical experiments presented, for example, by Gilbert and ocedal [7] proved that the PRP method is one of the best conjugate gradient methods, and this is the main motivation to consider it in (7) 3

4 On the other hand, the DY method alwa generates descent directions, and in [8] Dai established a remarable property for the DY conjugate gradient algorithm, relating the descent directions to the sufficient descent condition It is shown that if there exist constants γ and γ such that γ g γ for all, then for any p ( 0, ), there exists a constant c > 0 such that the sufficient descent condition gi di c gi holds for at least p indices i [ 0, ], where j denotes the largest integer j herefore, this property is the main reason we consider DY method in (7) It easy to see that: y g g g d g ( θ) + s θ = s (8) g g y s d d0 Supposing that is a descent direction ( (8) we can prove the following result = g ), then for the algorithm given by () and heorem Assume that α in algorithm () and (8) is determined by Wolfe line search (4) and (5) If 0< θ <, and then direction d + θ Proof Since 0< <, from (8) we get But, + + g + g 0 g s ( g y )( g s ), (9) given by (8) is a descent direction y g g g g d g g s g s = + + ( θ) + + θ + g g y g g g g+ + g + s + g + s g g g s y g = + + gs y g+ = g + + g + s gg + + g+ g + s gg gs > 0 by (5) and since 0, it follows that g s g 0 + herefore, from (9), it follows that g d + + 0, ie the direction d + is a descent one heorem Suppose that ( g+ y)( g+ s) 0 If 0< < then the direction d + given by (8) satisfies the sufficient descent condition g s + g+ d+ θ g + (0) Proof From (8) we have: g+ y g+ g+ g+ d+ = g+ + ( θ) g + s + θ g + s g g y s θ 4

5 g s ( g y )( g s ) = g + + θ + g ( θ) gg g s + θ g 0 + Observe that, since > 0 by (5) and since g + s = + gs <, then / g+ s > herefore, if 0< θ <, it follows that θ < / g+ s herefore g+ s θ > 0 proving the theorem o select the parameter θ we consider the following two possibilities In the first hybrid conjugate gradient algorithm the parameter θ is selected in such a manner that the conjugacy condition Hence, from yd + = 0 0 is satisfied at every iteration, independent on the line search yd + = after some algebra, using (8), we get: ( y g )( y s ) ( y g )( g g + + θ = ( y g+ )( ) g+ g In the second algorithm the parameter θ is selected in such a manner that the direction d + from (8) is the ewton direction, ie y g g g f ( x ) g g ( θ) + s θ = s () g g y s Having in view that f ( x ) s = y, from () we get: θ + ) ( y g+ s g+ ) g ( g+ y)( = g+ g ( g+ y)( ) Observe that the parameter θ given by () or (3) can be outside the interval [0,] However, in order to have a real convex combination in (7) the following rule is considered: PRP if θ 0, then set θ = 0 in (7), ie = ; if θ, then tae θ = in (7), ie DY = herefore, under this rule for properties of PRP and DY algorithms ) () (3) θ selection, the direction d + in (8) combines the 3 he ew Hybrid Algorithms (CCOMB, DOMB) n Step Initialization Select x0 R and the parameters 0< ρ σ < Compute f ( x 0) and g 0 Consider d 0 = g 0 and set the initial guess: α 0 = / g0 Step est for continuation of iterations If g 0 6, then stop Step 3 Line search Compute α > 0 satisfying the Wolfe line search condition (4) and (5) and update the variables x = x + + αd Compute f ( x + ), g + and s = x+ x, y = g g + Step 4 θ parameter computation If otherwise compute θ as follows: ( y g )( y s ) g g =, then set θ = 0, 5

6 CCOMB algorithm ( θ from Conjugacy Condition): ( y g+ )( ) ( y g+ )( g g θ = ( y g )( y s ) ( g g )( g g DOMB algorithm ( θ from ewton Direction): θ ) ) ( y g+ s g+ ) g ( g+ y)( = g+ g ( g+ y)( ) Step 5 conjugate gradient parameter computation If 0< θ <, then compute (7) If θ, then set = DY Step 6 Direction computation Compute is satisfied, then set α PRP If θ 0, then set = d = g + s + g+ g 0 g+, d = + g+ otherwise define d = α d / d, set = + and continue with step ) as in If the restart criterion of Powell (4) d + = Compute the initial guess It is well nown that if f is bounded along the direction then there exists a stepsize α satisfying the Wolfe line search conditions (4) and (5) In our algorithm when the Powell restart condition is satisfied, then we restart the algorithm with the negative gradient g + More sophisticated reasons for restarting the algorithms have been proposed in the literature [0], but we are interested in the performance of a conjugate gradient algorithm that uses this restart criterion, associated to a direction satisfying the conjugacy condition or is equal to the ewton direction Under reasonable assumptions, conditions (4), (5) and (4) are sufficient to prove the global convergence of the algorithm We consider this aspect in the next section he first trial of the step length crucially affects the practical behavior of the algorithm At every iteration the starting guess for the step α in the line search is computed as α d / d his selection was considered for the first time by Shanno and Phua in COMI [7] It is also considered in the pacages: SCG by Birgin and Martínez [5] and in SCALCG by Andrei [,3,4] 4 Convergence analis hroughout this section we assume that: S = x R n : f( x) f( x ) is bounded (i) he level set { 0 } (ii) In a neighborhood of S, the function f is continuously differentiable and its gradient is Lipschitz continuous, ie there exists a constant L > 0 such that f ( x) f( y) L x y, for all x, y Under these assumptions on f, there exists a constant Γ 0 such that f( x) Γ, for all x S In [] it is proved that for any conjugate gradient method with strong Wolfe line search the following general holds: Lemma Suppose that the assumptions (i) and (ii) hold and consider any conjugate gradient method () and (3), where d is a descent direction and α is obtained by the strong Wolfe line search If d 6

7 then =, (5) d lim inf g = 0 (6) For uniformly convex functions which satisfy the above assumptions we can prove that the norm of generated by (8) is bounded above hus, by lemma we have the following result d + heorem 3 Suppose that the assumptions (i) and (ii) hold Consider the algorithm () and (8), where is a descent direction and α is obtained by the strong Wolfe line search If for 0, that d + f ( x + α d ) f ( x ) ρα g d (7), g + d σ g d (8) s tends to zero and there exists the nonnegative constants η and η such g η s and g+ η s, (9) and the function f is a uniformly convex function, ie there exists a constant µ 0 such that for all x, y S then Proof From (0) it follows that ( f ( x) f( y)) ( x y) µ x y, (0) lim g = 0 () µ s ow, since 0 θ, from uniform convexity and (9) we have: g y g g g g y s But y Hence, L s, therefore + ΓL η + η s µ s d g s g+ y η s + η s µ s Γ+ +, η µ which implies that (5) is true herefore, by lemma we have (6), which for uniformly convex functions is equivalent to () Powell [5] showed that for general functions the PRP method is globally convergent if the steplengths s = x+ x tend to zero, ie s s is a condition of convergence For convergence of our algorithms from (9) we see that along the iterations, for, the gradient must be bounded as: η s g η s If the Powell condition is satisfied, ie s tends to zero, then s s ΓL η and therefore the norm of gradient can satisfy (9) 7

8 In the numerical experiments we observed that (9) is alwa satisfied in the last part of the iterations For general nonlinear functions the convergence analis of our algorithm exploits insights developed by Gilbert and ocedal [7], Dai and Liao [9] and that of Hager and Zhang [8] Global convergence proof of these new hybrid conjugate gradient algorithms is based on the Zoutendij condition combined with the analis showing that the sufficient descent condition holds and d is bounded Suppose that the level set L is bounded and the function f is bounded from below Lemma Assume that d is a descent direction and f satisfies the Lipschitz condition f ( x) f( x) L x x for all x on the line segment connecting x and x +, where L is a constant If the line search satisfies the second Wolfe condition (5), then σ gd α () L d Proof Subtracting Since g d from both sides of (5) and using the Lipschitz condition we have ( ) ( ) α (3) σ gd g+ g d L d d is a descent direction and σ <, () follows immediately from (3) heorem 4 Suppose that the assumptions (i) and (ii) holds, 0< θ, ( g+ y)( g+ s) 0, for every 0 there exists a positive constant ω, such that ( g s )/( y s ) ω > 0 and there exists the constants γ and Γ, such that for all, θ + γ g Γ hen for the computational scheme () and (8), where α is determined by the Wolfe line search (4) and (5), either g = 0 for some or liminf g = 0 (4) Proof By the Wolfe condition (5) we have: ( ) ( ) = g g s σ gs = ( σ) gs + By heorem, and the assumption θ( g+ s)/( ) ω, it follows that herefore, Combining (5) with (6) we get g d g s g g θ ω y s (5) g d ω g (6) σ ωα γ ( ) On the other hand y = g+ g L s Hence g y g y Γ L s With these, from (7) we get But, + + g y g g g g 8

9 g y g y ΓL s + ΓLD g g γ γ γ where D max { y z : y, z } On the other hand, herefore, +, = S is the diameter of the level set S ow, we can write g g + + Γ ( ) σ ωα γ ΓLD Γ + E (7) γ ( σ) ωα γ + + d g + s Γ+ ED (8) Since the level set L is bounded and the function f is bounded from below, using Lemma, from (4) it follows that ( gd) 0 < <, (9) = 0 d ie the Zoutendij condition holds herefore, from heorem using (9), the descent property yields: 4 4 γ g ( gd) = 0 d = 0 d = 0ω d <, which contradicts (8) Hence, γ = liminf g = 0 herefore, when 0< θ our hybrid conjugate gradient algorithms are globally convergent, meaning that either g = 0 for some or (4) holds Observe that in conditions of heorem the direction satisfies the sufficient descent condition independent on the line search d + 5 umerical experiments In this section we present the computational performance of a Fortran implementation of the CCOMB and DOMB algorithms on a set of 750 unconstrained optimization test problems he test problems are the unconstrained problems in the CUE [6] library, along with other large-scale optimization problems presented in [] We selected 75 large-scale unconstrained optimization problems in extended or generalized form Each problem is tested 0 times for a gradually increasing number of variables: n = 000, 000,, 0000 At the same time we present comparisons with other conjugate gradient algorithms, including the performance profiles of Dolan and Moré [4] All algorithms implement the Wolfe line search conditions with ρ = 0000 and σ = 09, and the same stopping criterion g 0 6, where is the maximum absolute component of a vector ALG he comparisons of algorithms are given in the following context Let f i and ALG f i be the optimal value found by ALG and ALG, for problem i =,, 750, respectively We say that, in the particular problem i, the performance of ALG was better than the performance of ALG if: 9

10 ALG ALG 3 fi fi <0 (30) and the number of iterations, or the number of function-gradient evaluations, or the CPU time of ALG was less than the number of iterations, or the number of function-gradient evaluations, or the CPU time corresponding to ALG, respectively All codes are written in double precision Fortran and compiled with f77 (default compiler settings) on an Intel Pentium 4, 8GHz worstation All these codes are authored by Andrei he performances of these algorithms have been evaluated using the profiles of Dolan and Moré [4] hat is, for each algorithm we plot the fraction of problems for which the algorithm is within a factor of the best CPU time he left side of these Figures gives the percentage of the test problems, out of 750, for which an algorithm is more performant; the right side gives the percentage of the test problems that were successfully solved by each of the algorithms Mainly, the right side represents a measure of an algorithm s robustness In the first set of numerical experiments we compare the performance of CCOMB to DOMB Figure shows the Dolan and Moré CPU performance profiles of CCOMB versus DOMB Fig Performance based on CPU time CCOMB versus DOMB Observe that CCOMB outperforms DOMB in the vast majority of problems Only 730 problems out 750 satisfy the criterion (30) Referring to the CPU time, CCOMB was better in 575 problems, in contrast with DOMB which solved only 7 problems in a better CPU time In the second set of numerical experiments we compare the performance of CCOMB to the PRP and DY conjugate gradient algorithms Figures and 3 show the Dolan and Moré CPU performance profiles of CCOMB versus PRP and CCOMB versus DY, respectively 0

11 Fig Performance based on CPU time CCOMB versus Pola-Ribière-Polya (PRP) Fig 3 Performance based on CPU time CCOMB versus Dai-Yuan (DY) When comparing CCOMB to PRP (Figure ), subject to the number of iterations, we see that CCOMB was better in 34 problems (ie it achieved the minimum number of iterations in 34 problems), PRP was better in 96 problems and they achieved the same number of iterations in 9 problems, etc Out of 750 problems, only for 7 problems the criteria (30) holds Similarly, in Figure 3 we see the number of problems for which CCOMB was better than DY Observe that the convex combination of PRP and DY, expressed as in (7), is far more successful than PRP or DY algorithms he third set of numerical experiments refers to the comparisons of CCOMB to hybrid conjugate gradient algorithms: hdy, hdyz, G, HuS, S and LS-CD Figures 4-9 presents the Dolan and Moré CPU performance profiles of these algorithms, as well as the

12 number of problems solved by each algorithms in minimum number of iterations, minimum number of function evaluations and minimum CPU time, respectively Fig 4 Performance based on CPU time CCOMB versus hybrid Dai-Yuan (hdy) Fig 5 Performance based on CPU time CCOMB versus hybrid Dai-Yuan (hdyz)

13 Fig 6 Performance based on CPU time CCOMB versus Gilbert-ocedal (G) Fig 7 Performance based on CPU time CCOMB versus Hu-Storey (HuS) 3

14 Fig 8 Performance based on CPU time CCOMB versus ouati-ahmed - Storey (S) Fig 9 Performance based on CPU time CCOMB versus Liu-Storey Conjugate Descent (LS-CD) From these Figures above we see that CCOMB is top performer Since these codes use the same Wolfe line search and the same stopping criterion they differ in their choice of the search direction Hence, among these conjugate gradient algorithms we considered here, CCOMB appears to generate the best search direction In the fourth set of numerical experiments we compare CCOMB to CG_DESCE conjugate gradient algorithm of Hager and Zhang [8] he computational scheme implemented in CG_DESCE is a modification of the Hestenes and Stiefel method which 4

15 satisfies the sufficient descent condition, independent of the accuracy of the line search he CG_DESCE code, authored by Hager and Zhang, contains the variant CG_DESCE (HZw) implementing the Wolfe line search and the variant CG_DESCE (HZaw) implementing an approximate Wolfe line search here are two main points associated to CG_DESCE Firstly, the scalar products are implemented using the loop unrolling of 6 depth 5 his is efficient for large-scale problems (over 0 variables) Secondly, the Wolfe line search is implemented using a very fine numerical interpretation of the first Wolfe condition (4) he Wolfe conditions implemented in CCOMB and CG_DESCE (HZw) can compute a solution with accuracy of the order of the square root of the machine epsilon Fig 0 Performance based on CPU time CCOMB versus CG_DESCE with Wolfe line search (HZw) Fig Performance based on CPU time CCOMB versus CG_DESCE with approximate Wolfe line search (HZaw) 5

16 In contrast, the approximate Wolfe line search implemented in CG_DESCE (HZaw) can compute the solution with accuracy of the order of machine epsilon Figures 0 and present the performance profile of these algorithms in comparison to CCOMB We see that CG_DESCE is more robust than CCOMB 5 Conclusion We now a large variety of conjugate gradient algorithms he nown hybrid conjugate gradient algorithms are based on projection of classical conjugate gradient algorithms In this paper we have proposed new hybrid conjugate gradient algorithms in which the famous DY parameter is computed as a convex combination of and, ie PRP DY = ( θ) + θ he parameter θ is computed in such a manner that the conjugacy condition is satisfied, or the corresponding direction in hybrid conjugate gradient algorithm is the ewton direction For uniformly convex functions if the stepsize approaches zero, the gradient is bounded in the sense that η s g η s and the line search satisfy the strong Wolfe conditions, then our hybrid conjugate gradient algorithms are globally convergent For general nonlinear functions if the parameter θ from definition is bounded, ie 0< θ <, then our hybrid conjugate gradient is globally PRP convergent he Dolan and Moré CPU performance profile of hybrid conjugate gradient algorithm based on conjugacy condition (CCOMB algorithm) is higher than the performance profile corresponding to the hybrid algorithm based to the ewton direction (DOMB algorithm) he performance profile of CCOMB algorithm was higher than those of the well established conjugate gradient algorithms (hdy, hdyz, G, HuS, S and LS-CD) for a set consisting of 750 unconstrained optimization test problems, some of them from CUE library Additionally the proposed hybrid conjugate gradient algorithm CCOMB is more robust than the PRP and DY conjugate gradient algorithms However, CCOMB algorithm is outperformed by CG_DESCE References [] Andrei, est functions for unconstrained optimization [] Andrei, Scaled conjugate gradient algorithms for unconstrained optimization Accepted: Computational Optimization and Applications, 006 [3] Andrei, Scaled memoryless BFGS preconditioned conjugate gradient algorithm for unconstrained optimization Accepted: Optimization Methods and Software, 006 [4] Andrei, A scaled BFGS preconditioned conjugate gradient algorithm for unconstrained optimization Accepted: Applied Mathematics Letters, 006 [5] E Birgin and JM Martínez, A spectral conjugate gradient method for unconstrained optimization, Applied Math and Optimization, 43, pp7-8, 00 [6] I Bongartz, AR Conn, IM Gould and PL oint, CUE: constrained and unconstrained testing environments, ACM rans Math Software,, pp3-60, 995 [7] YH Dai, Analis of conjugate gradient methods, PhD hesis, Institute of Computational Mathematics and Scientific/Engineering Computing, Chinese Academy of Science, 997 [8] YH Dai, ew properties of a nonlinear conjugate gradient method umer Math, 89 (00), pp83-98 [9] YH Dai and LZ Liao, ew conjugacy conditions and related nonlinear conjugate gradient methods Appl Math Optim, 43 (00), pp 87-0 [0] YH Dai, LZ Liao and Duan Li, On restart procedures for the conjugate gradient method umerical Algorithms 35 (004), pp s 6

17 [] YH Dai, Han, JY, Liu, GH, Sun, DF, Yin, X and Yuan, Y, Convergence properties of nonlinear conjugate gradient methods SIAM Journal on Optimization 0 (999), [] YH Dai and Y Yuan, A nonlinear conjugate gradient method with a strong global convergence property, SIAM J Optim, 0 (999), pp 77-8 [3] YH Dai and Y Yuan, An efficient hybrid conjugate gradient method for unconstrained optimization, Ann Oper Res, 03 (00), pp [4] ED Dolan and JJ Moré, Benchmaring optimization software with performance profiles, Math Programming, 9 (00), pp 0-3 [5] R Fletcher, Practical Methods of Optimization, vol : Unconstrained Optimization, John Wiley & Sons, ew Yor, 987 [6] R Fletcher and C Reeves, Function minimization by conjugate gradients, Comput J, 7 (964), pp49-54 [7] JC Gilbert and J ocedal, Global convergence properties of conjugate gradient methods for optimization, SIAM J Optim, (99), pp -4 [8] WW Hager and H Zhang, A new conjugate gradient method with guaranteed descent and an efficient line search, SIAM Journal on Optimization, 6 (005) 70-9 [9] WW Hager and H Zhang, A survey of nonlinear conjugate gradient methods Pacific journal of Optimization, (006), pp35-58 [0] MR Hestenes and EL Stiefel, Methods of conjugate gradients for solving linear stems, J Research at Bur Standards, 49 (95), pp [] YF Hu and C Storey, Global convergence result for conjugate gradient methods J Optim heory Appl, 7 (99), pp [] Y Liu, and C Storey, Efficient generalized conjugate gradient algorithms, Part : heory JOA, 69 (99), pp9-37 [3] E Pola and G Ribière, ote sur la convergence de directions conjuguée, Rev Francaise Informat Recherche Operationelle, 3e Année 6 (969), pp35-43 [4] B Polya, he conjugate gradient method in extreme problems USSR Comp Math Math Ph, 9 (969), pp94- [5] MJD Powell, Restart procedures of the conjugate gradient method Mathematical Programming, (977), pp4-54 [6] MJD Powell, onconvex minimization calculations and the conjugate gradient method in umerical Analis (Dundee, 983), Lecture otes in Mathematics, vol 066, Springer-Verlag, Berlin, 984, pp-4 [7] DF Shanno and KH Phua, Algorithm 500, Minimization of unconstrained multivariate functions, ACM rans on Math Soft,, pp87-94, 976 [8] D ouati-ahmed and C Storey, Efficient hybrid conjugate gradient techniques J Optim heory Appl, 64 (990), pp [9] YYuan, Analis on the conjugate gradient method, Optimization Methods and Software, (993), pp9-9 September 6, 007 Sent to Panos Pardalos 7

New Hybrid Conjugate Gradient Method as a Convex Combination of FR and PRP Methods

New Hybrid Conjugate Gradient Method as a Convex Combination of FR and PRP Methods Filomat 3:11 (216), 383 31 DOI 1.2298/FIL161183D Published by Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat New Hybrid Conjugate Gradient

More information

New Accelerated Conjugate Gradient Algorithms for Unconstrained Optimization

New Accelerated Conjugate Gradient Algorithms for Unconstrained Optimization ew Accelerated Conjugate Gradient Algorithms for Unconstrained Optimization eculai Andrei Research Institute for Informatics, Center for Advanced Modeling and Optimization, 8-0, Averescu Avenue, Bucharest,

More information

First Published on: 11 October 2006 To link to this article: DOI: / URL:

First Published on: 11 October 2006 To link to this article: DOI: / URL: his article was downloaded by:[universitetsbiblioteet i Bergen] [Universitetsbiblioteet i Bergen] On: 12 March 2007 Access Details: [subscription number 768372013] Publisher: aylor & Francis Informa Ltd

More information

New hybrid conjugate gradient methods with the generalized Wolfe line search

New hybrid conjugate gradient methods with the generalized Wolfe line search Xu and Kong SpringerPlus (016)5:881 DOI 10.1186/s40064-016-5-9 METHODOLOGY New hybrid conjugate gradient methods with the generalized Wolfe line search Open Access Xiao Xu * and Fan yu Kong *Correspondence:

More information

AN EIGENVALUE STUDY ON THE SUFFICIENT DESCENT PROPERTY OF A MODIFIED POLAK-RIBIÈRE-POLYAK CONJUGATE GRADIENT METHOD S.

AN EIGENVALUE STUDY ON THE SUFFICIENT DESCENT PROPERTY OF A MODIFIED POLAK-RIBIÈRE-POLYAK CONJUGATE GRADIENT METHOD S. Bull. Iranian Math. Soc. Vol. 40 (2014), No. 1, pp. 235 242 Online ISSN: 1735-8515 AN EIGENVALUE STUDY ON THE SUFFICIENT DESCENT PROPERTY OF A MODIFIED POLAK-RIBIÈRE-POLYAK CONJUGATE GRADIENT METHOD S.

More information

GLOBAL CONVERGENCE OF CONJUGATE GRADIENT METHODS WITHOUT LINE SEARCH

GLOBAL CONVERGENCE OF CONJUGATE GRADIENT METHODS WITHOUT LINE SEARCH GLOBAL CONVERGENCE OF CONJUGATE GRADIENT METHODS WITHOUT LINE SEARCH Jie Sun 1 Department of Decision Sciences National University of Singapore, Republic of Singapore Jiapu Zhang 2 Department of Mathematics

More information

Open Problems in Nonlinear Conjugate Gradient Algorithms for Unconstrained Optimization

Open Problems in Nonlinear Conjugate Gradient Algorithms for Unconstrained Optimization BULLETIN of the Malaysian Mathematical Sciences Society http://math.usm.my/bulletin Bull. Malays. Math. Sci. Soc. (2) 34(2) (2011), 319 330 Open Problems in Nonlinear Conjugate Gradient Algorithms for

More information

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications Weijun Zhou 28 October 20 Abstract A hybrid HS and PRP type conjugate gradient method for smooth

More information

Bulletin of the. Iranian Mathematical Society

Bulletin of the. Iranian Mathematical Society ISSN: 1017-060X (Print) ISSN: 1735-8515 (Online) Bulletin of the Iranian Mathematical Society Vol. 43 (2017), No. 7, pp. 2437 2448. Title: Extensions of the Hestenes Stiefel and Pola Ribière Polya conjugate

More information

Modification of the Armijo line search to satisfy the convergence properties of HS method

Modification of the Armijo line search to satisfy the convergence properties of HS method Université de Sfax Faculté des Sciences de Sfax Département de Mathématiques BP. 1171 Rte. Soukra 3000 Sfax Tunisia INTERNATIONAL CONFERENCE ON ADVANCES IN APPLIED MATHEMATICS 2014 Modification of the

More information

A Modified Hestenes-Stiefel Conjugate Gradient Method and Its Convergence

A Modified Hestenes-Stiefel Conjugate Gradient Method and Its Convergence Journal of Mathematical Research & Exposition Mar., 2010, Vol. 30, No. 2, pp. 297 308 DOI:10.3770/j.issn:1000-341X.2010.02.013 Http://jmre.dlut.edu.cn A Modified Hestenes-Stiefel Conjugate Gradient Method

More information

Global Convergence Properties of the HS Conjugate Gradient Method

Global Convergence Properties of the HS Conjugate Gradient Method Applied Mathematical Sciences, Vol. 7, 2013, no. 142, 7077-7091 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2013.311638 Global Convergence Properties of the HS Conjugate Gradient Method

More information

Convergence of a Two-parameter Family of Conjugate Gradient Methods with a Fixed Formula of Stepsize

Convergence of a Two-parameter Family of Conjugate Gradient Methods with a Fixed Formula of Stepsize Bol. Soc. Paran. Mat. (3s.) v. 00 0 (0000):????. c SPM ISSN-2175-1188 on line ISSN-00378712 in press SPM: www.spm.uem.br/bspm doi:10.5269/bspm.v38i6.35641 Convergence of a Two-parameter Family of Conjugate

More information

A Nonlinear Conjugate Gradient Algorithm with An Optimal Property and An Improved Wolfe Line Search

A Nonlinear Conjugate Gradient Algorithm with An Optimal Property and An Improved Wolfe Line Search A Nonlinear Conjugate Gradient Algorithm with An Optimal Property and An Improved Wolfe Line Search Yu-Hong Dai and Cai-Xia Kou State Key Laboratory of Scientific and Engineering Computing, Institute of

More information

Research Article A Descent Dai-Liao Conjugate Gradient Method Based on a Modified Secant Equation and Its Global Convergence

Research Article A Descent Dai-Liao Conjugate Gradient Method Based on a Modified Secant Equation and Its Global Convergence International Scholarly Research Networ ISRN Computational Mathematics Volume 2012, Article ID 435495, 8 pages doi:10.5402/2012/435495 Research Article A Descent Dai-Liao Conjugate Gradient Method Based

More information

An Efficient Modification of Nonlinear Conjugate Gradient Method

An Efficient Modification of Nonlinear Conjugate Gradient Method Malaysian Journal of Mathematical Sciences 10(S) March : 167-178 (2016) Special Issue: he 10th IM-G International Conference on Mathematics, Statistics and its Applications 2014 (ICMSA 2014) MALAYSIAN

More information

on descent spectral cg algorithm for training recurrent neural networks

on descent spectral cg algorithm for training recurrent neural networks Technical R e p o r t on descent spectral cg algorithm for training recurrent neural networs I.E. Livieris, D.G. Sotiropoulos 2 P. Pintelas,3 No. TR9-4 University of Patras Department of Mathematics GR-265

More information

Nonlinear conjugate gradient methods, Unconstrained optimization, Nonlinear

Nonlinear conjugate gradient methods, Unconstrained optimization, Nonlinear A SURVEY OF NONLINEAR CONJUGATE GRADIENT METHODS WILLIAM W. HAGER AND HONGCHAO ZHANG Abstract. This paper reviews the development of different versions of nonlinear conjugate gradient methods, with special

More information

A modified quadratic hybridization of Polak-Ribière-Polyak and Fletcher-Reeves conjugate gradient method for unconstrained optimization problems

A modified quadratic hybridization of Polak-Ribière-Polyak and Fletcher-Reeves conjugate gradient method for unconstrained optimization problems An International Journal of Optimization Control: Theories & Applications ISSN:246-0957 eissn:246-5703 Vol.7, No.2, pp.77-85 (207) http://doi.org/0.2/ijocta.0.207.00339 RESEARCH ARTICLE A modified quadratic

More information

NUMERICAL COMPARISON OF LINE SEARCH CRITERIA IN NONLINEAR CONJUGATE GRADIENT ALGORITHMS

NUMERICAL COMPARISON OF LINE SEARCH CRITERIA IN NONLINEAR CONJUGATE GRADIENT ALGORITHMS NUMERICAL COMPARISON OF LINE SEARCH CRITERIA IN NONLINEAR CONJUGATE GRADIENT ALGORITHMS Adeleke O. J. Department of Computer and Information Science/Mathematics Covenat University, Ota. Nigeria. Aderemi

More information

Step-size Estimation for Unconstrained Optimization Methods

Step-size Estimation for Unconstrained Optimization Methods Volume 24, N. 3, pp. 399 416, 2005 Copyright 2005 SBMAC ISSN 0101-8205 www.scielo.br/cam Step-size Estimation for Unconstrained Optimization Methods ZHEN-JUN SHI 1,2 and JIE SHEN 3 1 College of Operations

More information

Step lengths in BFGS method for monotone gradients

Step lengths in BFGS method for monotone gradients Noname manuscript No. (will be inserted by the editor) Step lengths in BFGS method for monotone gradients Yunda Dong Received: date / Accepted: date Abstract In this paper, we consider how to directly

More information

On Descent Spectral CG algorithms for Training Recurrent Neural Networks

On Descent Spectral CG algorithms for Training Recurrent Neural Networks 29 3th Panhellenic Conference on Informatics On Descent Spectral CG algorithms for Training Recurrent Neural Networs I.E. Livieris, D.G. Sotiropoulos and P. Pintelas Abstract In this paper, we evaluate

More information

January 29, Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes Stiefel 1 / 13

January 29, Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes Stiefel 1 / 13 Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière Hestenes Stiefel January 29, 2014 Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes

More information

and P RP k = gt k (g k? g k? ) kg k? k ; (.5) where kk is the Euclidean norm. This paper deals with another conjugate gradient method, the method of s

and P RP k = gt k (g k? g k? ) kg k? k ; (.5) where kk is the Euclidean norm. This paper deals with another conjugate gradient method, the method of s Global Convergence of the Method of Shortest Residuals Yu-hong Dai and Ya-xiang Yuan State Key Laboratory of Scientic and Engineering Computing, Institute of Computational Mathematics and Scientic/Engineering

More information

An Alternative Three-Term Conjugate Gradient Algorithm for Systems of Nonlinear Equations

An Alternative Three-Term Conjugate Gradient Algorithm for Systems of Nonlinear Equations International Journal of Mathematical Modelling & Computations Vol. 07, No. 02, Spring 2017, 145-157 An Alternative Three-Term Conjugate Gradient Algorithm for Systems of Nonlinear Equations L. Muhammad

More information

On the convergence properties of the modified Polak Ribiére Polyak method with the standard Armijo line search

On the convergence properties of the modified Polak Ribiére Polyak method with the standard Armijo line search ANZIAM J. 55 (E) pp.e79 E89, 2014 E79 On the convergence properties of the modified Polak Ribiére Polyak method with the standard Armijo line search Lijun Li 1 Weijun Zhou 2 (Received 21 May 2013; revised

More information

Adaptive two-point stepsize gradient algorithm

Adaptive two-point stepsize gradient algorithm Numerical Algorithms 27: 377 385, 2001. 2001 Kluwer Academic Publishers. Printed in the Netherlands. Adaptive two-point stepsize gradient algorithm Yu-Hong Dai and Hongchao Zhang State Key Laboratory of

More information

Research Article Nonlinear Conjugate Gradient Methods with Wolfe Type Line Search

Research Article Nonlinear Conjugate Gradient Methods with Wolfe Type Line Search Abstract and Applied Analysis Volume 013, Article ID 74815, 5 pages http://dx.doi.org/10.1155/013/74815 Research Article Nonlinear Conjugate Gradient Methods with Wolfe Type Line Search Yuan-Yuan Chen

More information

The Wolfe Epsilon Steepest Descent Algorithm

The Wolfe Epsilon Steepest Descent Algorithm Applied Mathematical Sciences, Vol. 8, 204, no. 55, 273-274 HIKARI Ltd, www.m-hiari.com http://dx.doi.org/0.2988/ams.204.4375 The Wolfe Epsilon Steepest Descent Algorithm Haima Degaichia Department of

More information

Preconditioned conjugate gradient algorithms with column scaling

Preconditioned conjugate gradient algorithms with column scaling Proceedings of the 47th IEEE Conference on Decision and Control Cancun, Mexico, Dec. 9-11, 28 Preconditioned conjugate gradient algorithms with column scaling R. Pytla Institute of Automatic Control and

More information

A family of derivative-free conjugate gradient methods for large-scale nonlinear systems of equations

A family of derivative-free conjugate gradient methods for large-scale nonlinear systems of equations Journal of Computational Applied Mathematics 224 (2009) 11 19 Contents lists available at ScienceDirect Journal of Computational Applied Mathematics journal homepage: www.elsevier.com/locate/cam A family

More information

A New Nonlinear Conjugate Gradient Coefficient. for Unconstrained Optimization

A New Nonlinear Conjugate Gradient Coefficient. for Unconstrained Optimization Applie Mathematical Sciences, Vol. 9, 05, no. 37, 83-8 HIKARI Lt, www.m-hiari.com http://x.oi.or/0.988/ams.05.4994 A New Nonlinear Conjuate Graient Coefficient for Unconstraine Optimization * Mohame Hamoa,

More information

Two new spectral conjugate gradient algorithms based on Hestenes Stiefel

Two new spectral conjugate gradient algorithms based on Hestenes Stiefel Research Article Two new spectral conjuate radient alorithms based on Hestenes Stiefel Journal of Alorithms & Computational Technoloy 207, Vol. (4) 345 352! The Author(s) 207 Reprints and permissions:

More information

Spectral gradient projection method for solving nonlinear monotone equations

Spectral gradient projection method for solving nonlinear monotone equations Journal of Computational and Applied Mathematics 196 (2006) 478 484 www.elsevier.com/locate/cam Spectral gradient projection method for solving nonlinear monotone equations Li Zhang, Weijun Zhou Department

More information

A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality constraints

A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality constraints Journal of Computational and Applied Mathematics 161 (003) 1 5 www.elsevier.com/locate/cam A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality

More information

Improved Damped Quasi-Newton Methods for Unconstrained Optimization

Improved Damped Quasi-Newton Methods for Unconstrained Optimization Improved Damped Quasi-Newton Methods for Unconstrained Optimization Mehiddin Al-Baali and Lucio Grandinetti August 2015 Abstract Recently, Al-Baali (2014) has extended the damped-technique in the modified

More information

2010 Mathematics Subject Classification: 90C30. , (1.2) where t. 0 is a step size, received from the line search, and the directions d

2010 Mathematics Subject Classification: 90C30. , (1.2) where t. 0 is a step size, received from the line search, and the directions d Journal of Applie Mathematics an Computation (JAMC), 8, (9), 366-378 http://wwwhillpublisheror/journal/jamc ISSN Online:576-645 ISSN Print:576-653 New Hbri Conjuate Graient Metho as A Convex Combination

More information

Research Article A New Conjugate Gradient Algorithm with Sufficient Descent Property for Unconstrained Optimization

Research Article A New Conjugate Gradient Algorithm with Sufficient Descent Property for Unconstrained Optimization Mathematical Problems in Engineering Volume 205, Article ID 352524, 8 pages http://dx.doi.org/0.55/205/352524 Research Article A New Conjugate Gradient Algorithm with Sufficient Descent Property for Unconstrained

More information

A derivative-free nonmonotone line search and its application to the spectral residual method

A derivative-free nonmonotone line search and its application to the spectral residual method IMA Journal of Numerical Analysis (2009) 29, 814 825 doi:10.1093/imanum/drn019 Advance Access publication on November 14, 2008 A derivative-free nonmonotone line search and its application to the spectral

More information

Globally convergent three-term conjugate gradient projection methods for solving nonlinear monotone equations

Globally convergent three-term conjugate gradient projection methods for solving nonlinear monotone equations Arab. J. Math. (2018) 7:289 301 https://doi.org/10.1007/s40065-018-0206-8 Arabian Journal of Mathematics Mompati S. Koorapetse P. Kaelo Globally convergent three-term conjugate gradient projection methods

More information

Conjugate gradient methods based on secant conditions that generate descent search directions for unconstrained optimization

Conjugate gradient methods based on secant conditions that generate descent search directions for unconstrained optimization Conjugate gradient methods based on secant conditions that generate descent search directions for unconstrained optimization Yasushi Narushima and Hiroshi Yabe September 28, 2011 Abstract Conjugate gradient

More information

THE RELATIONSHIPS BETWEEN CG, BFGS, AND TWO LIMITED-MEMORY ALGORITHMS

THE RELATIONSHIPS BETWEEN CG, BFGS, AND TWO LIMITED-MEMORY ALGORITHMS Furman University Electronic Journal of Undergraduate Mathematics Volume 12, 5 20, 2007 HE RELAIONSHIPS BEWEEN CG, BFGS, AND WO LIMIED-MEMORY ALGORIHMS ZHIWEI (ONY) QIN Abstract. For the solution of linear

More information

Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method. 1 Introduction

Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method. 1 Introduction ISSN 1749-3889 (print), 1749-3897 (online) International Journal of Nonlinear Science Vol.11(2011) No.2,pp.153-158 Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method Yigui Ou, Jun Zhang

More information

230 L. HEI if ρ k is satisfactory enough, and to reduce it by a constant fraction (say, ahalf): k+1 = fi 2 k (0 <fi 2 < 1); (1.7) in the case ρ k is n

230 L. HEI if ρ k is satisfactory enough, and to reduce it by a constant fraction (say, ahalf): k+1 = fi 2 k (0 <fi 2 < 1); (1.7) in the case ρ k is n Journal of Computational Mathematics, Vol.21, No.2, 2003, 229 236. A SELF-ADAPTIVE TRUST REGION ALGORITHM Λ1) Long Hei y (Institute of Computational Mathematics and Scientific/Engineering Computing, Academy

More information

Math 164: Optimization Barzilai-Borwein Method

Math 164: Optimization Barzilai-Borwein Method Math 164: Optimization Barzilai-Borwein Method Instructor: Wotao Yin Department of Mathematics, UCLA Spring 2015 online discussions on piazza.com Main features of the Barzilai-Borwein (BB) method The BB

More information

HT14 A COMPARISON OF DIFFERENT PARAMETER ESTIMATION TECHNIQUES FOR THE IDENTIFICATION OF THERMAL CONDUCTIVITY COMPONENTS OF ORTHOTROPIC SOLIDS

HT14 A COMPARISON OF DIFFERENT PARAMETER ESTIMATION TECHNIQUES FOR THE IDENTIFICATION OF THERMAL CONDUCTIVITY COMPONENTS OF ORTHOTROPIC SOLIDS Inverse Problems in Engineering: heory and Practice rd Int. Conference on Inverse Problems in Engineering June -8, 999, Port Ludlow, WA, USA H4 A COMPARISON OF DIFFEREN PARAMEER ESIMAION ECHNIQUES FOR

More information

New Inexact Line Search Method for Unconstrained Optimization 1,2

New Inexact Line Search Method for Unconstrained Optimization 1,2 journal of optimization theory and applications: Vol. 127, No. 2, pp. 425 446, November 2005 ( 2005) DOI: 10.1007/s10957-005-6553-6 New Inexact Line Search Method for Unconstrained Optimization 1,2 Z.

More information

Quasi-Newton Methods

Quasi-Newton Methods Quasi-Newton Methods Werner C. Rheinboldt These are excerpts of material relating to the boos [OR00 and [Rhe98 and of write-ups prepared for courses held at the University of Pittsburgh. Some further references

More information

Downloaded 12/02/13 to Redistribution subject to SIAM license or copyright; see

Downloaded 12/02/13 to Redistribution subject to SIAM license or copyright; see SIAM J. OPTIM. Vol. 23, No. 4, pp. 2150 2168 c 2013 Society for Industrial and Applied Mathematics THE LIMITED MEMORY CONJUGATE GRADIENT METHOD WILLIAM W. HAGER AND HONGCHAO ZHANG Abstract. In theory,

More information

1 Numerical optimization

1 Numerical optimization Contents 1 Numerical optimization 5 1.1 Optimization of single-variable functions............ 5 1.1.1 Golden Section Search................... 6 1.1. Fibonacci Search...................... 8 1. Algorithms

More information

Line Search Methods for Unconstrained Optimisation

Line Search Methods for Unconstrained Optimisation Line Search Methods for Unconstrained Optimisation Lecture 8, Numerical Linear Algebra and Optimisation Oxford University Computing Laboratory, MT 2007 Dr Raphael Hauser (hauser@comlab.ox.ac.uk) The Generic

More information

Gradient method based on epsilon algorithm for large-scale nonlinearoptimization

Gradient method based on epsilon algorithm for large-scale nonlinearoptimization ISSN 1746-7233, England, UK World Journal of Modelling and Simulation Vol. 4 (2008) No. 1, pp. 64-68 Gradient method based on epsilon algorithm for large-scale nonlinearoptimization Jianliang Li, Lian

More information

1 Numerical optimization

1 Numerical optimization Contents Numerical optimization 5. Optimization of single-variable functions.............................. 5.. Golden Section Search..................................... 6.. Fibonacci Search........................................

More information

A new initial search direction for nonlinear conjugate gradient method

A new initial search direction for nonlinear conjugate gradient method International Journal of Mathematis Researh. ISSN 0976-5840 Volume 6, Number 2 (2014), pp. 183 190 International Researh Publiation House http://www.irphouse.om A new initial searh diretion for nonlinear

More information

GRADIENT METHODS FOR LARGE-SCALE NONLINEAR OPTIMIZATION

GRADIENT METHODS FOR LARGE-SCALE NONLINEAR OPTIMIZATION GRADIENT METHODS FOR LARGE-SCALE NONLINEAR OPTIMIZATION By HONGCHAO ZHANG A DISSERTATION PRESENTED TO THE GRADUATE SCHOOL OF THE UNIVERSITY OF FLORIDA IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE

More information

NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM

NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM JIAYI GUO AND A.S. LEWIS Abstract. The popular BFGS quasi-newton minimization algorithm under reasonable conditions converges globally on smooth

More information

Unconstrained optimization

Unconstrained optimization Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout

More information

A line-search algorithm inspired by the adaptive cubic regularization framework, with a worst-case complexity O(ɛ 3/2 ).

A line-search algorithm inspired by the adaptive cubic regularization framework, with a worst-case complexity O(ɛ 3/2 ). A line-search algorithm inspired by the adaptive cubic regularization framewor, with a worst-case complexity Oɛ 3/. E. Bergou Y. Diouane S. Gratton June 16, 017 Abstract Adaptive regularized framewor using

More information

Gradient methods exploiting spectral properties

Gradient methods exploiting spectral properties Noname manuscript No. (will be inserted by the editor) Gradient methods exploiting spectral properties Yaui Huang Yu-Hong Dai Xin-Wei Liu Received: date / Accepted: date Abstract A new stepsize is derived

More information

Statistics 580 Optimization Methods

Statistics 580 Optimization Methods Statistics 580 Optimization Methods Introduction Let fx be a given real-valued function on R p. The general optimization problem is to find an x ɛ R p at which fx attain a maximum or a minimum. It is of

More information

A class of Smoothing Method for Linear Second-Order Cone Programming

A class of Smoothing Method for Linear Second-Order Cone Programming Columbia International Publishing Journal of Advanced Computing (13) 1: 9-4 doi:1776/jac1313 Research Article A class of Smoothing Method for Linear Second-Order Cone Programming Zhuqing Gui *, Zhibin

More information

A DIMENSION REDUCING CONIC METHOD FOR UNCONSTRAINED OPTIMIZATION

A DIMENSION REDUCING CONIC METHOD FOR UNCONSTRAINED OPTIMIZATION 1 A DIMENSION REDUCING CONIC METHOD FOR UNCONSTRAINED OPTIMIZATION G E MANOUSSAKIS, T N GRAPSA and C A BOTSARIS Department of Mathematics, University of Patras, GR 26110 Patras, Greece e-mail :gemini@mathupatrasgr,

More information

A Conjugate Gradient Method with Inexact. Line Search for Unconstrained Optimization

A Conjugate Gradient Method with Inexact. Line Search for Unconstrained Optimization Applie Mathematical Sciences, Vol. 9, 5, no. 37, 83-83 HIKARI Lt, www.m-hiari.com http://x.oi.or/.988/ams.5.4995 A Conuate Graient Metho with Inexact Line Search for Unconstraine Optimization * Mohame

More information

SOLVING SYSTEMS OF NONLINEAR EQUATIONS USING A GLOBALLY CONVERGENT OPTIMIZATION ALGORITHM

SOLVING SYSTEMS OF NONLINEAR EQUATIONS USING A GLOBALLY CONVERGENT OPTIMIZATION ALGORITHM Transaction on Evolutionary Algorithm and Nonlinear Optimization ISSN: 2229-87 Online Publication, June 2012 www.pcoglobal.com/gjto.htm NG-O21/GJTO SOLVING SYSTEMS OF NONLINEAR EQUATIONS USING A GLOBALLY

More information

ON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS

ON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS ON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS Anders FORSGREN Tove ODLAND Technical Report TRITA-MAT-203-OS-03 Department of Mathematics KTH Royal

More information

Chapter 4. Unconstrained optimization

Chapter 4. Unconstrained optimization Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file

More information

DENSE INITIALIZATIONS FOR LIMITED-MEMORY QUASI-NEWTON METHODS

DENSE INITIALIZATIONS FOR LIMITED-MEMORY QUASI-NEWTON METHODS DENSE INITIALIZATIONS FOR LIMITED-MEMORY QUASI-NEWTON METHODS by Johannes Brust, Oleg Burdaov, Jennifer B. Erway, and Roummel F. Marcia Technical Report 07-, Department of Mathematics and Statistics, Wae

More information

A new class of Conjugate Gradient Methods with extended Nonmonotone Line Search

A new class of Conjugate Gradient Methods with extended Nonmonotone Line Search Appl. Math. Inf. Sci. 6, No. 1S, 147S-154S (2012) 147 Applied Mathematics & Information Sciences An International Journal A new class of Conjugate Gradient Methods with extended Nonmonotone Line Search

More information

A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES

A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES IJMMS 25:6 2001) 397 409 PII. S0161171201002290 http://ijmms.hindawi.com Hindawi Publishing Corp. A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES

More information

Modification of the Wolfe line search rules to satisfy the descent condition in the Polak-Ribière-Polyak conjugate gradient method

Modification of the Wolfe line search rules to satisfy the descent condition in the Polak-Ribière-Polyak conjugate gradient method Laboratoire d Arithmétique, Calcul formel et d Optimisation UMR CNRS 6090 Modification of the Wolfe line search rules to satisfy the descent condition in the Polak-Ribière-Polyak conjugate gradient method

More information

Research Article Finding Global Minima with a Filled Function Approach for Non-Smooth Global Optimization

Research Article Finding Global Minima with a Filled Function Approach for Non-Smooth Global Optimization Hindawi Publishing Corporation Discrete Dynamics in Nature and Society Volume 00, Article ID 843609, 0 pages doi:0.55/00/843609 Research Article Finding Global Minima with a Filled Function Approach for

More information

Higher-Order Methods

Higher-Order Methods Higher-Order Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. PCMI, July 2016 Stephen Wright (UW-Madison) Higher-Order Methods PCMI, July 2016 1 / 25 Smooth

More information

Nonmonotonic back-tracking trust region interior point algorithm for linear constrained optimization

Nonmonotonic back-tracking trust region interior point algorithm for linear constrained optimization Journal of Computational and Applied Mathematics 155 (2003) 285 305 www.elsevier.com/locate/cam Nonmonotonic bac-tracing trust region interior point algorithm for linear constrained optimization Detong

More information

Modification of Non-Linear Constrained Optimization. Ban A. Metras Dep. of Statistics College of Computer Sciences and Mathematics

Modification of Non-Linear Constrained Optimization. Ban A. Metras Dep. of Statistics College of Computer Sciences and Mathematics Raf. J. Of Comp. & Math s., Vol., No., 004 Modification of Non-Linear Constrained Optimization Ban A. Metras Dep. of Statistics College of Computer Sciences and Mathematics Received: 3/5/00 Accepted: /0/00

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week

More information

FALL 2018 MATH 4211/6211 Optimization Homework 4

FALL 2018 MATH 4211/6211 Optimization Homework 4 FALL 2018 MATH 4211/6211 Optimization Homework 4 This homework assignment is open to textbook, reference books, slides, and online resources, excluding any direct solution to the problem (such as solution

More information

Convergence rate of SGD

Convergence rate of SGD Convergence rate of SGD heorem: (see Nemirovski et al 09 from readings) Let f be a strongly convex stochastic function Assume gradient of f is Lipschitz continuous and bounded hen, for step sizes: he expected

More information

Modification of the Wolfe line search rules to satisfy the descent condition in the Polak-Ribière-Polyak conjugate gradient method

Modification of the Wolfe line search rules to satisfy the descent condition in the Polak-Ribière-Polyak conjugate gradient method Laboratoire d Arithmétique, Calcul formel et d Optimisation UMR CNRS 6090 Modification of the Wolfe line search rules to satisfy the descent condition in the Polak-Ribière-Polyak conjugate gradient method

More information

Acceleration Method for Convex Optimization over the Fixed Point Set of a Nonexpansive Mapping

Acceleration Method for Convex Optimization over the Fixed Point Set of a Nonexpansive Mapping Noname manuscript No. will be inserted by the editor) Acceleration Method for Convex Optimization over the Fixed Point Set of a Nonexpansive Mapping Hideaki Iiduka Received: date / Accepted: date Abstract

More information

Unconstrained Multivariate Optimization

Unconstrained Multivariate Optimization Unconstrained Multivariate Optimization Multivariate optimization means optimization of a scalar function of a several variables: and has the general form: y = () min ( ) where () is a nonlinear scalar-valued

More information

On the acceleration of augmented Lagrangian method for linearly constrained optimization

On the acceleration of augmented Lagrangian method for linearly constrained optimization On the acceleration of augmented Lagrangian method for linearly constrained optimization Bingsheng He and Xiaoming Yuan October, 2 Abstract. The classical augmented Lagrangian method (ALM plays a fundamental

More information

17 Solution of Nonlinear Systems

17 Solution of Nonlinear Systems 17 Solution of Nonlinear Systems We now discuss the solution of systems of nonlinear equations. An important ingredient will be the multivariate Taylor theorem. Theorem 17.1 Let D = {x 1, x 2,..., x m

More information

Journal of Computational and Applied Mathematics. Notes on the Dai Yuan Yuan modified spectral gradient method

Journal of Computational and Applied Mathematics. Notes on the Dai Yuan Yuan modified spectral gradient method Journal of Computational Applied Mathematics 234 (200) 2986 2992 Contents lists available at ScienceDirect Journal of Computational Applied Mathematics journal homepage: wwwelseviercom/locate/cam Notes

More information

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term;

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term; Chapter 2 Gradient Methods The gradient method forms the foundation of all of the schemes studied in this book. We will provide several complementary perspectives on this algorithm that highlight the many

More information

Multipoint secant and interpolation methods with nonmonotone line search for solving systems of nonlinear equations

Multipoint secant and interpolation methods with nonmonotone line search for solving systems of nonlinear equations Multipoint secant and interpolation methods with nonmonotone line search for solving systems of nonlinear equations Oleg Burdakov a,, Ahmad Kamandi b a Department of Mathematics, Linköping University,

More information

Residual iterative schemes for largescale linear systems

Residual iterative schemes for largescale linear systems Universidad Central de Venezuela Facultad de Ciencias Escuela de Computación Lecturas en Ciencias de la Computación ISSN 1316-6239 Residual iterative schemes for largescale linear systems William La Cruz

More information

A Novel of Step Size Selection Procedures. for Steepest Descent Method

A Novel of Step Size Selection Procedures. for Steepest Descent Method Applied Mathematical Sciences, Vol. 6, 0, no. 5, 507 58 A Novel of Step Size Selection Procedures for Steepest Descent Method Goh Khang Wen, Mustafa Mamat, Ismail bin Mohd, 3 Yosza Dasril Department of

More information

A line-search algorithm inspired by the adaptive cubic regularization framework with a worst-case complexity O(ɛ 3/2 )

A line-search algorithm inspired by the adaptive cubic regularization framework with a worst-case complexity O(ɛ 3/2 ) A line-search algorithm inspired by the adaptive cubic regularization framewor with a worst-case complexity Oɛ 3/ E. Bergou Y. Diouane S. Gratton December 4, 017 Abstract Adaptive regularized framewor

More information

Solutions and Notes to Selected Problems In: Numerical Optimzation by Jorge Nocedal and Stephen J. Wright.

Solutions and Notes to Selected Problems In: Numerical Optimzation by Jorge Nocedal and Stephen J. Wright. Solutions and Notes to Selected Problems In: Numerical Optimzation by Jorge Nocedal and Stephen J. Wright. John L. Weatherwax July 7, 2010 wax@alum.mit.edu 1 Chapter 5 (Conjugate Gradient Methods) Notes

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428

More information

A Trust Region Algorithm Model With Radius Bounded Below for Minimization of Locally Lipschitzian Functions

A Trust Region Algorithm Model With Radius Bounded Below for Minimization of Locally Lipschitzian Functions The First International Symposium on Optimization and Systems Biology (OSB 07) Beijing, China, August 8 10, 2007 Copyright 2007 ORSC & APORC pp. 405 411 A Trust Region Algorithm Model With Radius Bounded

More information

Introduction to Nonlinear Optimization Paul J. Atzberger

Introduction to Nonlinear Optimization Paul J. Atzberger Introduction to Nonlinear Optimization Paul J. Atzberger Comments should be sent to: atzberg@math.ucsb.edu Introduction We shall discuss in these notes a brief introduction to nonlinear optimization concepts,

More information

On Lagrange multipliers of trust-region subproblems

On Lagrange multipliers of trust-region subproblems On Lagrange multipliers of trust-region subproblems Ladislav Lukšan, Ctirad Matonoha, Jan Vlček Institute of Computer Science AS CR, Prague Programy a algoritmy numerické matematiky 14 1.- 6. června 2008

More information

A NOVEL FILLED FUNCTION METHOD FOR GLOBAL OPTIMIZATION. 1. Introduction Consider the following unconstrained programming problem:

A NOVEL FILLED FUNCTION METHOD FOR GLOBAL OPTIMIZATION. 1. Introduction Consider the following unconstrained programming problem: J. Korean Math. Soc. 47, No. 6, pp. 53 67 DOI.434/JKMS..47.6.53 A NOVEL FILLED FUNCTION METHOD FOR GLOBAL OPTIMIZATION Youjiang Lin, Yongjian Yang, and Liansheng Zhang Abstract. This paper considers the

More information

OPER 627: Nonlinear Optimization Lecture 14: Mid-term Review

OPER 627: Nonlinear Optimization Lecture 14: Mid-term Review OPER 627: Nonlinear Optimization Lecture 14: Mid-term Review Department of Statistical Sciences and Operations Research Virginia Commonwealth University Oct 16, 2013 (Lecture 14) Nonlinear Optimization

More information

Jae Heon Yun and Yu Du Han

Jae Heon Yun and Yu Du Han Bull. Korean Math. Soc. 39 (2002), No. 3, pp. 495 509 MODIFIED INCOMPLETE CHOLESKY FACTORIZATION PRECONDITIONERS FOR A SYMMETRIC POSITIVE DEFINITE MATRIX Jae Heon Yun and Yu Du Han Abstract. We propose

More information

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09 Numerical Optimization 1 Working Horse in Computer Vision Variational Methods Shape Analysis Machine Learning Markov Random Fields Geometry Common denominator: optimization problems 2 Overview of Methods

More information

Steepest descent algorithm. Conjugate gradient training algorithm. Steepest descent algorithm. Remember previous examples

Steepest descent algorithm. Conjugate gradient training algorithm. Steepest descent algorithm. Remember previous examples Conjugate gradient training algorithm Steepest descent algorithm So far: Heuristic improvements to gradient descent (momentum Steepest descent training algorithm Can we do better? Definitions: weight vector

More information

Global convergence of a regularized factorized quasi-newton method for nonlinear least squares problems

Global convergence of a regularized factorized quasi-newton method for nonlinear least squares problems Volume 29, N. 2, pp. 195 214, 2010 Copyright 2010 SBMAC ISSN 0101-8205 www.scielo.br/cam Global convergence of a regularized factorized quasi-newton method for nonlinear least squares problems WEIJUN ZHOU

More information