New hybrid conjugate gradient algorithms for unconstrained optimization
|
|
- Oswald Bryan
- 5 years ago
- Views:
Transcription
1 ew hybrid conjugate gradient algorithms for unconstrained optimization eculai Andrei Research Institute for Informatics, Center for Advanced Modeling and Optimization, 8-0, Averescu Avenue, Bucharest, Romania, Abstract ew hybrid conjugate gradient algorithms are proposed and analyzed In these hybrid algorithms the famous parameter is computed as a convex combination of the Pola-Ribière-Polya and Dai-Yuan conjugate gradient algorithms In one hybrid algorithm the parameter in convex combination is computed in such a way that the conjugacy condition is satisfied, independent of the line search In the other, the parameter in convex combination is computed in such a way that the conjugate gradient direction is the ewton direction he algorithm uses the standard Wolfe line search conditions umerical comparisons with conjugate gradient algorithms using a set of 750 unconstrained optimization problems, some of them from the CUE library, show that the hybrid computational scheme based on conjugacy condition outperform the nown hybrid conjugate gradient algorithms MSC: 49M07, 49M0, 90C06, 65K Keywords: Unconstrained optimization, hybrid conjugate gradient method, conjugacy condition, numerical comparisons Introduction Let us consider the nonlinear unconstrained optimization problem n min f( x): x R, () { } n where f : R R is a continuously differentiable function, bounded from below For solving this problem, starting from an initial guess x0 generates a sequence { x } as R n, a nonlinear conjugate gradient method, x = x + + αd, () where α > 0 is obtained by line search, and the directions d are generated as d+ = g+ + s, d0 = g0 (3) In (3) is nown as the conjugate gradient parameter, s = x+ x and g = f( x) Consider the Euclidean norm and define y = g+ g he line search in the conjugate gradient algorithms often is based on the standard Wolfe conditions: f ( x + α d ) f ( x ) ρα g d (4), g+ d σ gd, (5) where d is a descent direction and 0< ρ σ < Plenty of conjugate gradient methods are nown, and an excellent survey of these methods, with a special attention on their global convergence, is given by Hager and Zhang [9] Different conjugate gradient algorithms correspond to different choices for the scalar parameter Some of these methods as
2 Fletcher and Reeves (FR) [6], Dai and Yuan (DY) [] and Conjugate Descent (CD) proposed by Fletcher [5]: FR g+ g+ DY g+ g+ CD g+ g+ =, =, =, g g gs have strong convergence properties, but they may have modest practical performance due to jamming On the other hand, the methods of Pola Ribière [3] and Polya (PRP) [4], Hestenes and Stiefel (HS) [0] or Liu and Storey (LS) []: PRP g+ y HS g+ y LS g+ y =,, =, = g g gs in general may not be convergent, but they often have better computational performances In this paper we focus on hybrid conjugate gradient methods hese methods are combinations of different conjugate gradient algorithms, mainly they being proposed to avoid the jamming phenomenon One of the first hybrid conjugate gradient algorithms has been introduced by ouati-ahmed and Storey [8], where the parameter is computed as: PRP g+ y PRP FR =, if 0, S g = FR g + =, otherwise g he PRP method has a built-in restart feature that directly addresses to jamming Indeed, when the step s is small, then the factor y in the numerator of tends to zero herefore, descent direction PRP becomes small and the search direction g + PRP d + is very close to the steepest Hence, when the iterations jam, the method of ouati-ahmed and Storey uses the PRP computational scheme Another hybrid conjugate gradient method was given by Hu and Storey [], where is: { { }} = max 0, min, HuS PRP FR in (3) As above, when the method of Hu and Storey is jamming, then the PRP method is used instead he combination between LS and CD conjugate gradient methods leads to the following hybrid method: { { }} = max 0, min, LS CD LS CD he CD method of Fletcher [5] is very close to FR method With an exact line search, CD method is identical to FR Similarly, for an exact line search, LS method is also identical to PRP herefore, the hybrid LS-CD method with an exact line search has similar performances with the hybrid method of Hu and Storey Gilbert and ocedal [7] suggested a combination between PRP and FR methods as: Since FR { { }} G max FR, min PRP, FR = is alwa nonnegative, it follows that G can be negative he method of Gilbert and ocedal has the same advantage of avoiding jamming Using the standard Wolfe line search, the DY method alwa generates descent directions and if the gradient is Lipschitz continuously the method is global convergent In an effort to improve their algorithm, Dai and Yuan [3] combined their algorithm with other conjugate gradient algorithm, proposing the following two hybrid methods: { c { }} hdy max DY, min HS, DY =,
3 { { }} = max 0, min,, hdyz HS DY where c = ( σ ) /( + σ ) For the standard Wolfe conditions (4) and (5), under the Lipschitz continuity of the gradient, Dai and Yuan [3] established the global convergence of these hybrid computational schemes In this paper we propose another hybrid conjugate gradient as a convex combination of PRP and DY conjugate gradient algorithms We selected these two methods to combine in a hybrid conjugate gradient algorithm because PRP has good computational properties, on one side, and DY has strong convergence properties, on the other side Often PRP method performs better in practice than DY and we speculate this in order to have a good practical conjugate algorithm he structure of the paper is as follows In section we introduce our hybrid conjugate gradient algorithm and prove that it generates descent directions satisfying in some conditions the sufficient descent condition Section 3 presents the algorithms and in section 4 we show its convergence analis In section 5 some numerical experiments and performance profiles of Dolan-Moré [4] corresponding to this new hybrid conjugate gradient algorithm and some other conjugate gradient algorithms are presented he performance profiles corresponding to a set of 750 unconstrained optimization problems in the CUE test problem library [6], as well as some other unconstrained optimization problems presented in [] show that this hybrid conjugate gradient algorithm outperform the nown hybrid conjugate gradient algorithms ew hybrid conjugate gradient algorithms he iterates x, x, x, the stepsize directions where d 0 of our algorithm are computed by means of the recurrence () where α > 0 is determined according to the Wolfe conditions (4) and (5), and the are generated by the rule: d = g + s, d0 = g0, (6) + + = ( θ ) + θ PRP DY g y g g = ( θ ) + g g θ (7) and θ is a scalar parameter satisfying 0 θ, which follows to be determined Observe PRP DY that if θ = 0, then =, and if θ =, then = On the other hand, if PRP DY 0< θ <, then is a convex combination of and Referring to the PRP method, Pola and Ribière [3] proved that when function f is strongly convex and the line search is exact, then the PRP method is global convergent In an effort to understand the behavior of the PRP method, Powell [5] showed that if the step length s = x x approaches to zero, the line search is exact and the gradient f ( x) is + Lipschitz continuous, then the PRP method is globally convergent Additionally, assuming that the search direction is a descent direction, Yuan [9] established the global convergence of the PRP method for strongly convex functions and a Wolfe line search For general nonlinear functions the convergence of the PRP method is uncertain Powell [6] gave a 3 dimensional example, in which the function to be minimized is not strongly convex, showing that even with an exact line search, the PRP method may not converge to a stationary point Later on Dai [7] presented another example this time with a strongly convex function for which the PRP method fails to generate a descent direction herefore, theoretically the convergence of the PRP method is limited to strongly convex functions For general nonlinear functions the convergence of the PRP method is established under restrictive conditions (Lipschitz continuity, exact line search and the stepsize tends to zero) However, the numerical experiments presented, for example, by Gilbert and ocedal [7] proved that the PRP method is one of the best conjugate gradient methods, and this is the main motivation to consider it in (7) 3
4 On the other hand, the DY method alwa generates descent directions, and in [8] Dai established a remarable property for the DY conjugate gradient algorithm, relating the descent directions to the sufficient descent condition It is shown that if there exist constants γ and γ such that γ g γ for all, then for any p ( 0, ), there exists a constant c > 0 such that the sufficient descent condition gi di c gi holds for at least p indices i [ 0, ], where j denotes the largest integer j herefore, this property is the main reason we consider DY method in (7) It easy to see that: y g g g d g ( θ) + s θ = s (8) g g y s d d0 Supposing that is a descent direction ( (8) we can prove the following result = g ), then for the algorithm given by () and heorem Assume that α in algorithm () and (8) is determined by Wolfe line search (4) and (5) If 0< θ <, and then direction d + θ Proof Since 0< <, from (8) we get But, + + g + g 0 g s ( g y )( g s ), (9) given by (8) is a descent direction y g g g g d g g s g s = + + ( θ) + + θ + g g y g g g g+ + g + s + g + s g g g s y g = + + gs y g+ = g + + g + s gg + + g+ g + s gg gs > 0 by (5) and since 0, it follows that g s g 0 + herefore, from (9), it follows that g d + + 0, ie the direction d + is a descent one heorem Suppose that ( g+ y)( g+ s) 0 If 0< < then the direction d + given by (8) satisfies the sufficient descent condition g s + g+ d+ θ g + (0) Proof From (8) we have: g+ y g+ g+ g+ d+ = g+ + ( θ) g + s + θ g + s g g y s θ 4
5 g s ( g y )( g s ) = g + + θ + g ( θ) gg g s + θ g 0 + Observe that, since > 0 by (5) and since g + s = + gs <, then / g+ s > herefore, if 0< θ <, it follows that θ < / g+ s herefore g+ s θ > 0 proving the theorem o select the parameter θ we consider the following two possibilities In the first hybrid conjugate gradient algorithm the parameter θ is selected in such a manner that the conjugacy condition Hence, from yd + = 0 0 is satisfied at every iteration, independent on the line search yd + = after some algebra, using (8), we get: ( y g )( y s ) ( y g )( g g + + θ = ( y g+ )( ) g+ g In the second algorithm the parameter θ is selected in such a manner that the direction d + from (8) is the ewton direction, ie y g g g f ( x ) g g ( θ) + s θ = s () g g y s Having in view that f ( x ) s = y, from () we get: θ + ) ( y g+ s g+ ) g ( g+ y)( = g+ g ( g+ y)( ) Observe that the parameter θ given by () or (3) can be outside the interval [0,] However, in order to have a real convex combination in (7) the following rule is considered: PRP if θ 0, then set θ = 0 in (7), ie = ; if θ, then tae θ = in (7), ie DY = herefore, under this rule for properties of PRP and DY algorithms ) () (3) θ selection, the direction d + in (8) combines the 3 he ew Hybrid Algorithms (CCOMB, DOMB) n Step Initialization Select x0 R and the parameters 0< ρ σ < Compute f ( x 0) and g 0 Consider d 0 = g 0 and set the initial guess: α 0 = / g0 Step est for continuation of iterations If g 0 6, then stop Step 3 Line search Compute α > 0 satisfying the Wolfe line search condition (4) and (5) and update the variables x = x + + αd Compute f ( x + ), g + and s = x+ x, y = g g + Step 4 θ parameter computation If otherwise compute θ as follows: ( y g )( y s ) g g =, then set θ = 0, 5
6 CCOMB algorithm ( θ from Conjugacy Condition): ( y g+ )( ) ( y g+ )( g g θ = ( y g )( y s ) ( g g )( g g DOMB algorithm ( θ from ewton Direction): θ ) ) ( y g+ s g+ ) g ( g+ y)( = g+ g ( g+ y)( ) Step 5 conjugate gradient parameter computation If 0< θ <, then compute (7) If θ, then set = DY Step 6 Direction computation Compute is satisfied, then set α PRP If θ 0, then set = d = g + s + g+ g 0 g+, d = + g+ otherwise define d = α d / d, set = + and continue with step ) as in If the restart criterion of Powell (4) d + = Compute the initial guess It is well nown that if f is bounded along the direction then there exists a stepsize α satisfying the Wolfe line search conditions (4) and (5) In our algorithm when the Powell restart condition is satisfied, then we restart the algorithm with the negative gradient g + More sophisticated reasons for restarting the algorithms have been proposed in the literature [0], but we are interested in the performance of a conjugate gradient algorithm that uses this restart criterion, associated to a direction satisfying the conjugacy condition or is equal to the ewton direction Under reasonable assumptions, conditions (4), (5) and (4) are sufficient to prove the global convergence of the algorithm We consider this aspect in the next section he first trial of the step length crucially affects the practical behavior of the algorithm At every iteration the starting guess for the step α in the line search is computed as α d / d his selection was considered for the first time by Shanno and Phua in COMI [7] It is also considered in the pacages: SCG by Birgin and Martínez [5] and in SCALCG by Andrei [,3,4] 4 Convergence analis hroughout this section we assume that: S = x R n : f( x) f( x ) is bounded (i) he level set { 0 } (ii) In a neighborhood of S, the function f is continuously differentiable and its gradient is Lipschitz continuous, ie there exists a constant L > 0 such that f ( x) f( y) L x y, for all x, y Under these assumptions on f, there exists a constant Γ 0 such that f( x) Γ, for all x S In [] it is proved that for any conjugate gradient method with strong Wolfe line search the following general holds: Lemma Suppose that the assumptions (i) and (ii) hold and consider any conjugate gradient method () and (3), where d is a descent direction and α is obtained by the strong Wolfe line search If d 6
7 then =, (5) d lim inf g = 0 (6) For uniformly convex functions which satisfy the above assumptions we can prove that the norm of generated by (8) is bounded above hus, by lemma we have the following result d + heorem 3 Suppose that the assumptions (i) and (ii) hold Consider the algorithm () and (8), where is a descent direction and α is obtained by the strong Wolfe line search If for 0, that d + f ( x + α d ) f ( x ) ρα g d (7), g + d σ g d (8) s tends to zero and there exists the nonnegative constants η and η such g η s and g+ η s, (9) and the function f is a uniformly convex function, ie there exists a constant µ 0 such that for all x, y S then Proof From (0) it follows that ( f ( x) f( y)) ( x y) µ x y, (0) lim g = 0 () µ s ow, since 0 θ, from uniform convexity and (9) we have: g y g g g g y s But y Hence, L s, therefore + ΓL η + η s µ s d g s g+ y η s + η s µ s Γ+ +, η µ which implies that (5) is true herefore, by lemma we have (6), which for uniformly convex functions is equivalent to () Powell [5] showed that for general functions the PRP method is globally convergent if the steplengths s = x+ x tend to zero, ie s s is a condition of convergence For convergence of our algorithms from (9) we see that along the iterations, for, the gradient must be bounded as: η s g η s If the Powell condition is satisfied, ie s tends to zero, then s s ΓL η and therefore the norm of gradient can satisfy (9) 7
8 In the numerical experiments we observed that (9) is alwa satisfied in the last part of the iterations For general nonlinear functions the convergence analis of our algorithm exploits insights developed by Gilbert and ocedal [7], Dai and Liao [9] and that of Hager and Zhang [8] Global convergence proof of these new hybrid conjugate gradient algorithms is based on the Zoutendij condition combined with the analis showing that the sufficient descent condition holds and d is bounded Suppose that the level set L is bounded and the function f is bounded from below Lemma Assume that d is a descent direction and f satisfies the Lipschitz condition f ( x) f( x) L x x for all x on the line segment connecting x and x +, where L is a constant If the line search satisfies the second Wolfe condition (5), then σ gd α () L d Proof Subtracting Since g d from both sides of (5) and using the Lipschitz condition we have ( ) ( ) α (3) σ gd g+ g d L d d is a descent direction and σ <, () follows immediately from (3) heorem 4 Suppose that the assumptions (i) and (ii) holds, 0< θ, ( g+ y)( g+ s) 0, for every 0 there exists a positive constant ω, such that ( g s )/( y s ) ω > 0 and there exists the constants γ and Γ, such that for all, θ + γ g Γ hen for the computational scheme () and (8), where α is determined by the Wolfe line search (4) and (5), either g = 0 for some or liminf g = 0 (4) Proof By the Wolfe condition (5) we have: ( ) ( ) = g g s σ gs = ( σ) gs + By heorem, and the assumption θ( g+ s)/( ) ω, it follows that herefore, Combining (5) with (6) we get g d g s g g θ ω y s (5) g d ω g (6) σ ωα γ ( ) On the other hand y = g+ g L s Hence g y g y Γ L s With these, from (7) we get But, + + g y g g g g 8
9 g y g y ΓL s + ΓLD g g γ γ γ where D max { y z : y, z } On the other hand, herefore, +, = S is the diameter of the level set S ow, we can write g g + + Γ ( ) σ ωα γ ΓLD Γ + E (7) γ ( σ) ωα γ + + d g + s Γ+ ED (8) Since the level set L is bounded and the function f is bounded from below, using Lemma, from (4) it follows that ( gd) 0 < <, (9) = 0 d ie the Zoutendij condition holds herefore, from heorem using (9), the descent property yields: 4 4 γ g ( gd) = 0 d = 0 d = 0ω d <, which contradicts (8) Hence, γ = liminf g = 0 herefore, when 0< θ our hybrid conjugate gradient algorithms are globally convergent, meaning that either g = 0 for some or (4) holds Observe that in conditions of heorem the direction satisfies the sufficient descent condition independent on the line search d + 5 umerical experiments In this section we present the computational performance of a Fortran implementation of the CCOMB and DOMB algorithms on a set of 750 unconstrained optimization test problems he test problems are the unconstrained problems in the CUE [6] library, along with other large-scale optimization problems presented in [] We selected 75 large-scale unconstrained optimization problems in extended or generalized form Each problem is tested 0 times for a gradually increasing number of variables: n = 000, 000,, 0000 At the same time we present comparisons with other conjugate gradient algorithms, including the performance profiles of Dolan and Moré [4] All algorithms implement the Wolfe line search conditions with ρ = 0000 and σ = 09, and the same stopping criterion g 0 6, where is the maximum absolute component of a vector ALG he comparisons of algorithms are given in the following context Let f i and ALG f i be the optimal value found by ALG and ALG, for problem i =,, 750, respectively We say that, in the particular problem i, the performance of ALG was better than the performance of ALG if: 9
10 ALG ALG 3 fi fi <0 (30) and the number of iterations, or the number of function-gradient evaluations, or the CPU time of ALG was less than the number of iterations, or the number of function-gradient evaluations, or the CPU time corresponding to ALG, respectively All codes are written in double precision Fortran and compiled with f77 (default compiler settings) on an Intel Pentium 4, 8GHz worstation All these codes are authored by Andrei he performances of these algorithms have been evaluated using the profiles of Dolan and Moré [4] hat is, for each algorithm we plot the fraction of problems for which the algorithm is within a factor of the best CPU time he left side of these Figures gives the percentage of the test problems, out of 750, for which an algorithm is more performant; the right side gives the percentage of the test problems that were successfully solved by each of the algorithms Mainly, the right side represents a measure of an algorithm s robustness In the first set of numerical experiments we compare the performance of CCOMB to DOMB Figure shows the Dolan and Moré CPU performance profiles of CCOMB versus DOMB Fig Performance based on CPU time CCOMB versus DOMB Observe that CCOMB outperforms DOMB in the vast majority of problems Only 730 problems out 750 satisfy the criterion (30) Referring to the CPU time, CCOMB was better in 575 problems, in contrast with DOMB which solved only 7 problems in a better CPU time In the second set of numerical experiments we compare the performance of CCOMB to the PRP and DY conjugate gradient algorithms Figures and 3 show the Dolan and Moré CPU performance profiles of CCOMB versus PRP and CCOMB versus DY, respectively 0
11 Fig Performance based on CPU time CCOMB versus Pola-Ribière-Polya (PRP) Fig 3 Performance based on CPU time CCOMB versus Dai-Yuan (DY) When comparing CCOMB to PRP (Figure ), subject to the number of iterations, we see that CCOMB was better in 34 problems (ie it achieved the minimum number of iterations in 34 problems), PRP was better in 96 problems and they achieved the same number of iterations in 9 problems, etc Out of 750 problems, only for 7 problems the criteria (30) holds Similarly, in Figure 3 we see the number of problems for which CCOMB was better than DY Observe that the convex combination of PRP and DY, expressed as in (7), is far more successful than PRP or DY algorithms he third set of numerical experiments refers to the comparisons of CCOMB to hybrid conjugate gradient algorithms: hdy, hdyz, G, HuS, S and LS-CD Figures 4-9 presents the Dolan and Moré CPU performance profiles of these algorithms, as well as the
12 number of problems solved by each algorithms in minimum number of iterations, minimum number of function evaluations and minimum CPU time, respectively Fig 4 Performance based on CPU time CCOMB versus hybrid Dai-Yuan (hdy) Fig 5 Performance based on CPU time CCOMB versus hybrid Dai-Yuan (hdyz)
13 Fig 6 Performance based on CPU time CCOMB versus Gilbert-ocedal (G) Fig 7 Performance based on CPU time CCOMB versus Hu-Storey (HuS) 3
14 Fig 8 Performance based on CPU time CCOMB versus ouati-ahmed - Storey (S) Fig 9 Performance based on CPU time CCOMB versus Liu-Storey Conjugate Descent (LS-CD) From these Figures above we see that CCOMB is top performer Since these codes use the same Wolfe line search and the same stopping criterion they differ in their choice of the search direction Hence, among these conjugate gradient algorithms we considered here, CCOMB appears to generate the best search direction In the fourth set of numerical experiments we compare CCOMB to CG_DESCE conjugate gradient algorithm of Hager and Zhang [8] he computational scheme implemented in CG_DESCE is a modification of the Hestenes and Stiefel method which 4
15 satisfies the sufficient descent condition, independent of the accuracy of the line search he CG_DESCE code, authored by Hager and Zhang, contains the variant CG_DESCE (HZw) implementing the Wolfe line search and the variant CG_DESCE (HZaw) implementing an approximate Wolfe line search here are two main points associated to CG_DESCE Firstly, the scalar products are implemented using the loop unrolling of 6 depth 5 his is efficient for large-scale problems (over 0 variables) Secondly, the Wolfe line search is implemented using a very fine numerical interpretation of the first Wolfe condition (4) he Wolfe conditions implemented in CCOMB and CG_DESCE (HZw) can compute a solution with accuracy of the order of the square root of the machine epsilon Fig 0 Performance based on CPU time CCOMB versus CG_DESCE with Wolfe line search (HZw) Fig Performance based on CPU time CCOMB versus CG_DESCE with approximate Wolfe line search (HZaw) 5
16 In contrast, the approximate Wolfe line search implemented in CG_DESCE (HZaw) can compute the solution with accuracy of the order of machine epsilon Figures 0 and present the performance profile of these algorithms in comparison to CCOMB We see that CG_DESCE is more robust than CCOMB 5 Conclusion We now a large variety of conjugate gradient algorithms he nown hybrid conjugate gradient algorithms are based on projection of classical conjugate gradient algorithms In this paper we have proposed new hybrid conjugate gradient algorithms in which the famous DY parameter is computed as a convex combination of and, ie PRP DY = ( θ) + θ he parameter θ is computed in such a manner that the conjugacy condition is satisfied, or the corresponding direction in hybrid conjugate gradient algorithm is the ewton direction For uniformly convex functions if the stepsize approaches zero, the gradient is bounded in the sense that η s g η s and the line search satisfy the strong Wolfe conditions, then our hybrid conjugate gradient algorithms are globally convergent For general nonlinear functions if the parameter θ from definition is bounded, ie 0< θ <, then our hybrid conjugate gradient is globally PRP convergent he Dolan and Moré CPU performance profile of hybrid conjugate gradient algorithm based on conjugacy condition (CCOMB algorithm) is higher than the performance profile corresponding to the hybrid algorithm based to the ewton direction (DOMB algorithm) he performance profile of CCOMB algorithm was higher than those of the well established conjugate gradient algorithms (hdy, hdyz, G, HuS, S and LS-CD) for a set consisting of 750 unconstrained optimization test problems, some of them from CUE library Additionally the proposed hybrid conjugate gradient algorithm CCOMB is more robust than the PRP and DY conjugate gradient algorithms However, CCOMB algorithm is outperformed by CG_DESCE References [] Andrei, est functions for unconstrained optimization [] Andrei, Scaled conjugate gradient algorithms for unconstrained optimization Accepted: Computational Optimization and Applications, 006 [3] Andrei, Scaled memoryless BFGS preconditioned conjugate gradient algorithm for unconstrained optimization Accepted: Optimization Methods and Software, 006 [4] Andrei, A scaled BFGS preconditioned conjugate gradient algorithm for unconstrained optimization Accepted: Applied Mathematics Letters, 006 [5] E Birgin and JM Martínez, A spectral conjugate gradient method for unconstrained optimization, Applied Math and Optimization, 43, pp7-8, 00 [6] I Bongartz, AR Conn, IM Gould and PL oint, CUE: constrained and unconstrained testing environments, ACM rans Math Software,, pp3-60, 995 [7] YH Dai, Analis of conjugate gradient methods, PhD hesis, Institute of Computational Mathematics and Scientific/Engineering Computing, Chinese Academy of Science, 997 [8] YH Dai, ew properties of a nonlinear conjugate gradient method umer Math, 89 (00), pp83-98 [9] YH Dai and LZ Liao, ew conjugacy conditions and related nonlinear conjugate gradient methods Appl Math Optim, 43 (00), pp 87-0 [0] YH Dai, LZ Liao and Duan Li, On restart procedures for the conjugate gradient method umerical Algorithms 35 (004), pp s 6
17 [] YH Dai, Han, JY, Liu, GH, Sun, DF, Yin, X and Yuan, Y, Convergence properties of nonlinear conjugate gradient methods SIAM Journal on Optimization 0 (999), [] YH Dai and Y Yuan, A nonlinear conjugate gradient method with a strong global convergence property, SIAM J Optim, 0 (999), pp 77-8 [3] YH Dai and Y Yuan, An efficient hybrid conjugate gradient method for unconstrained optimization, Ann Oper Res, 03 (00), pp [4] ED Dolan and JJ Moré, Benchmaring optimization software with performance profiles, Math Programming, 9 (00), pp 0-3 [5] R Fletcher, Practical Methods of Optimization, vol : Unconstrained Optimization, John Wiley & Sons, ew Yor, 987 [6] R Fletcher and C Reeves, Function minimization by conjugate gradients, Comput J, 7 (964), pp49-54 [7] JC Gilbert and J ocedal, Global convergence properties of conjugate gradient methods for optimization, SIAM J Optim, (99), pp -4 [8] WW Hager and H Zhang, A new conjugate gradient method with guaranteed descent and an efficient line search, SIAM Journal on Optimization, 6 (005) 70-9 [9] WW Hager and H Zhang, A survey of nonlinear conjugate gradient methods Pacific journal of Optimization, (006), pp35-58 [0] MR Hestenes and EL Stiefel, Methods of conjugate gradients for solving linear stems, J Research at Bur Standards, 49 (95), pp [] YF Hu and C Storey, Global convergence result for conjugate gradient methods J Optim heory Appl, 7 (99), pp [] Y Liu, and C Storey, Efficient generalized conjugate gradient algorithms, Part : heory JOA, 69 (99), pp9-37 [3] E Pola and G Ribière, ote sur la convergence de directions conjuguée, Rev Francaise Informat Recherche Operationelle, 3e Année 6 (969), pp35-43 [4] B Polya, he conjugate gradient method in extreme problems USSR Comp Math Math Ph, 9 (969), pp94- [5] MJD Powell, Restart procedures of the conjugate gradient method Mathematical Programming, (977), pp4-54 [6] MJD Powell, onconvex minimization calculations and the conjugate gradient method in umerical Analis (Dundee, 983), Lecture otes in Mathematics, vol 066, Springer-Verlag, Berlin, 984, pp-4 [7] DF Shanno and KH Phua, Algorithm 500, Minimization of unconstrained multivariate functions, ACM rans on Math Soft,, pp87-94, 976 [8] D ouati-ahmed and C Storey, Efficient hybrid conjugate gradient techniques J Optim heory Appl, 64 (990), pp [9] YYuan, Analis on the conjugate gradient method, Optimization Methods and Software, (993), pp9-9 September 6, 007 Sent to Panos Pardalos 7
New Hybrid Conjugate Gradient Method as a Convex Combination of FR and PRP Methods
Filomat 3:11 (216), 383 31 DOI 1.2298/FIL161183D Published by Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat New Hybrid Conjugate Gradient
More informationNew Accelerated Conjugate Gradient Algorithms for Unconstrained Optimization
ew Accelerated Conjugate Gradient Algorithms for Unconstrained Optimization eculai Andrei Research Institute for Informatics, Center for Advanced Modeling and Optimization, 8-0, Averescu Avenue, Bucharest,
More informationFirst Published on: 11 October 2006 To link to this article: DOI: / URL:
his article was downloaded by:[universitetsbiblioteet i Bergen] [Universitetsbiblioteet i Bergen] On: 12 March 2007 Access Details: [subscription number 768372013] Publisher: aylor & Francis Informa Ltd
More informationNew hybrid conjugate gradient methods with the generalized Wolfe line search
Xu and Kong SpringerPlus (016)5:881 DOI 10.1186/s40064-016-5-9 METHODOLOGY New hybrid conjugate gradient methods with the generalized Wolfe line search Open Access Xiao Xu * and Fan yu Kong *Correspondence:
More informationAN EIGENVALUE STUDY ON THE SUFFICIENT DESCENT PROPERTY OF A MODIFIED POLAK-RIBIÈRE-POLYAK CONJUGATE GRADIENT METHOD S.
Bull. Iranian Math. Soc. Vol. 40 (2014), No. 1, pp. 235 242 Online ISSN: 1735-8515 AN EIGENVALUE STUDY ON THE SUFFICIENT DESCENT PROPERTY OF A MODIFIED POLAK-RIBIÈRE-POLYAK CONJUGATE GRADIENT METHOD S.
More informationGLOBAL CONVERGENCE OF CONJUGATE GRADIENT METHODS WITHOUT LINE SEARCH
GLOBAL CONVERGENCE OF CONJUGATE GRADIENT METHODS WITHOUT LINE SEARCH Jie Sun 1 Department of Decision Sciences National University of Singapore, Republic of Singapore Jiapu Zhang 2 Department of Mathematics
More informationOpen Problems in Nonlinear Conjugate Gradient Algorithms for Unconstrained Optimization
BULLETIN of the Malaysian Mathematical Sciences Society http://math.usm.my/bulletin Bull. Malays. Math. Sci. Soc. (2) 34(2) (2011), 319 330 Open Problems in Nonlinear Conjugate Gradient Algorithms for
More informationA globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications
A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications Weijun Zhou 28 October 20 Abstract A hybrid HS and PRP type conjugate gradient method for smooth
More informationBulletin of the. Iranian Mathematical Society
ISSN: 1017-060X (Print) ISSN: 1735-8515 (Online) Bulletin of the Iranian Mathematical Society Vol. 43 (2017), No. 7, pp. 2437 2448. Title: Extensions of the Hestenes Stiefel and Pola Ribière Polya conjugate
More informationModification of the Armijo line search to satisfy the convergence properties of HS method
Université de Sfax Faculté des Sciences de Sfax Département de Mathématiques BP. 1171 Rte. Soukra 3000 Sfax Tunisia INTERNATIONAL CONFERENCE ON ADVANCES IN APPLIED MATHEMATICS 2014 Modification of the
More informationA Modified Hestenes-Stiefel Conjugate Gradient Method and Its Convergence
Journal of Mathematical Research & Exposition Mar., 2010, Vol. 30, No. 2, pp. 297 308 DOI:10.3770/j.issn:1000-341X.2010.02.013 Http://jmre.dlut.edu.cn A Modified Hestenes-Stiefel Conjugate Gradient Method
More informationGlobal Convergence Properties of the HS Conjugate Gradient Method
Applied Mathematical Sciences, Vol. 7, 2013, no. 142, 7077-7091 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2013.311638 Global Convergence Properties of the HS Conjugate Gradient Method
More informationConvergence of a Two-parameter Family of Conjugate Gradient Methods with a Fixed Formula of Stepsize
Bol. Soc. Paran. Mat. (3s.) v. 00 0 (0000):????. c SPM ISSN-2175-1188 on line ISSN-00378712 in press SPM: www.spm.uem.br/bspm doi:10.5269/bspm.v38i6.35641 Convergence of a Two-parameter Family of Conjugate
More informationA Nonlinear Conjugate Gradient Algorithm with An Optimal Property and An Improved Wolfe Line Search
A Nonlinear Conjugate Gradient Algorithm with An Optimal Property and An Improved Wolfe Line Search Yu-Hong Dai and Cai-Xia Kou State Key Laboratory of Scientific and Engineering Computing, Institute of
More informationResearch Article A Descent Dai-Liao Conjugate Gradient Method Based on a Modified Secant Equation and Its Global Convergence
International Scholarly Research Networ ISRN Computational Mathematics Volume 2012, Article ID 435495, 8 pages doi:10.5402/2012/435495 Research Article A Descent Dai-Liao Conjugate Gradient Method Based
More informationAn Efficient Modification of Nonlinear Conjugate Gradient Method
Malaysian Journal of Mathematical Sciences 10(S) March : 167-178 (2016) Special Issue: he 10th IM-G International Conference on Mathematics, Statistics and its Applications 2014 (ICMSA 2014) MALAYSIAN
More informationon descent spectral cg algorithm for training recurrent neural networks
Technical R e p o r t on descent spectral cg algorithm for training recurrent neural networs I.E. Livieris, D.G. Sotiropoulos 2 P. Pintelas,3 No. TR9-4 University of Patras Department of Mathematics GR-265
More informationNonlinear conjugate gradient methods, Unconstrained optimization, Nonlinear
A SURVEY OF NONLINEAR CONJUGATE GRADIENT METHODS WILLIAM W. HAGER AND HONGCHAO ZHANG Abstract. This paper reviews the development of different versions of nonlinear conjugate gradient methods, with special
More informationA modified quadratic hybridization of Polak-Ribière-Polyak and Fletcher-Reeves conjugate gradient method for unconstrained optimization problems
An International Journal of Optimization Control: Theories & Applications ISSN:246-0957 eissn:246-5703 Vol.7, No.2, pp.77-85 (207) http://doi.org/0.2/ijocta.0.207.00339 RESEARCH ARTICLE A modified quadratic
More informationNUMERICAL COMPARISON OF LINE SEARCH CRITERIA IN NONLINEAR CONJUGATE GRADIENT ALGORITHMS
NUMERICAL COMPARISON OF LINE SEARCH CRITERIA IN NONLINEAR CONJUGATE GRADIENT ALGORITHMS Adeleke O. J. Department of Computer and Information Science/Mathematics Covenat University, Ota. Nigeria. Aderemi
More informationStep-size Estimation for Unconstrained Optimization Methods
Volume 24, N. 3, pp. 399 416, 2005 Copyright 2005 SBMAC ISSN 0101-8205 www.scielo.br/cam Step-size Estimation for Unconstrained Optimization Methods ZHEN-JUN SHI 1,2 and JIE SHEN 3 1 College of Operations
More informationStep lengths in BFGS method for monotone gradients
Noname manuscript No. (will be inserted by the editor) Step lengths in BFGS method for monotone gradients Yunda Dong Received: date / Accepted: date Abstract In this paper, we consider how to directly
More informationOn Descent Spectral CG algorithms for Training Recurrent Neural Networks
29 3th Panhellenic Conference on Informatics On Descent Spectral CG algorithms for Training Recurrent Neural Networs I.E. Livieris, D.G. Sotiropoulos and P. Pintelas Abstract In this paper, we evaluate
More informationJanuary 29, Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes Stiefel 1 / 13
Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière Hestenes Stiefel January 29, 2014 Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes
More informationand P RP k = gt k (g k? g k? ) kg k? k ; (.5) where kk is the Euclidean norm. This paper deals with another conjugate gradient method, the method of s
Global Convergence of the Method of Shortest Residuals Yu-hong Dai and Ya-xiang Yuan State Key Laboratory of Scientic and Engineering Computing, Institute of Computational Mathematics and Scientic/Engineering
More informationAn Alternative Three-Term Conjugate Gradient Algorithm for Systems of Nonlinear Equations
International Journal of Mathematical Modelling & Computations Vol. 07, No. 02, Spring 2017, 145-157 An Alternative Three-Term Conjugate Gradient Algorithm for Systems of Nonlinear Equations L. Muhammad
More informationOn the convergence properties of the modified Polak Ribiére Polyak method with the standard Armijo line search
ANZIAM J. 55 (E) pp.e79 E89, 2014 E79 On the convergence properties of the modified Polak Ribiére Polyak method with the standard Armijo line search Lijun Li 1 Weijun Zhou 2 (Received 21 May 2013; revised
More informationAdaptive two-point stepsize gradient algorithm
Numerical Algorithms 27: 377 385, 2001. 2001 Kluwer Academic Publishers. Printed in the Netherlands. Adaptive two-point stepsize gradient algorithm Yu-Hong Dai and Hongchao Zhang State Key Laboratory of
More informationResearch Article Nonlinear Conjugate Gradient Methods with Wolfe Type Line Search
Abstract and Applied Analysis Volume 013, Article ID 74815, 5 pages http://dx.doi.org/10.1155/013/74815 Research Article Nonlinear Conjugate Gradient Methods with Wolfe Type Line Search Yuan-Yuan Chen
More informationThe Wolfe Epsilon Steepest Descent Algorithm
Applied Mathematical Sciences, Vol. 8, 204, no. 55, 273-274 HIKARI Ltd, www.m-hiari.com http://dx.doi.org/0.2988/ams.204.4375 The Wolfe Epsilon Steepest Descent Algorithm Haima Degaichia Department of
More informationPreconditioned conjugate gradient algorithms with column scaling
Proceedings of the 47th IEEE Conference on Decision and Control Cancun, Mexico, Dec. 9-11, 28 Preconditioned conjugate gradient algorithms with column scaling R. Pytla Institute of Automatic Control and
More informationA family of derivative-free conjugate gradient methods for large-scale nonlinear systems of equations
Journal of Computational Applied Mathematics 224 (2009) 11 19 Contents lists available at ScienceDirect Journal of Computational Applied Mathematics journal homepage: www.elsevier.com/locate/cam A family
More informationA New Nonlinear Conjugate Gradient Coefficient. for Unconstrained Optimization
Applie Mathematical Sciences, Vol. 9, 05, no. 37, 83-8 HIKARI Lt, www.m-hiari.com http://x.oi.or/0.988/ams.05.4994 A New Nonlinear Conjuate Graient Coefficient for Unconstraine Optimization * Mohame Hamoa,
More informationTwo new spectral conjugate gradient algorithms based on Hestenes Stiefel
Research Article Two new spectral conjuate radient alorithms based on Hestenes Stiefel Journal of Alorithms & Computational Technoloy 207, Vol. (4) 345 352! The Author(s) 207 Reprints and permissions:
More informationSpectral gradient projection method for solving nonlinear monotone equations
Journal of Computational and Applied Mathematics 196 (2006) 478 484 www.elsevier.com/locate/cam Spectral gradient projection method for solving nonlinear monotone equations Li Zhang, Weijun Zhou Department
More informationA new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality constraints
Journal of Computational and Applied Mathematics 161 (003) 1 5 www.elsevier.com/locate/cam A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality
More informationImproved Damped Quasi-Newton Methods for Unconstrained Optimization
Improved Damped Quasi-Newton Methods for Unconstrained Optimization Mehiddin Al-Baali and Lucio Grandinetti August 2015 Abstract Recently, Al-Baali (2014) has extended the damped-technique in the modified
More information2010 Mathematics Subject Classification: 90C30. , (1.2) where t. 0 is a step size, received from the line search, and the directions d
Journal of Applie Mathematics an Computation (JAMC), 8, (9), 366-378 http://wwwhillpublisheror/journal/jamc ISSN Online:576-645 ISSN Print:576-653 New Hbri Conjuate Graient Metho as A Convex Combination
More informationResearch Article A New Conjugate Gradient Algorithm with Sufficient Descent Property for Unconstrained Optimization
Mathematical Problems in Engineering Volume 205, Article ID 352524, 8 pages http://dx.doi.org/0.55/205/352524 Research Article A New Conjugate Gradient Algorithm with Sufficient Descent Property for Unconstrained
More informationA derivative-free nonmonotone line search and its application to the spectral residual method
IMA Journal of Numerical Analysis (2009) 29, 814 825 doi:10.1093/imanum/drn019 Advance Access publication on November 14, 2008 A derivative-free nonmonotone line search and its application to the spectral
More informationGlobally convergent three-term conjugate gradient projection methods for solving nonlinear monotone equations
Arab. J. Math. (2018) 7:289 301 https://doi.org/10.1007/s40065-018-0206-8 Arabian Journal of Mathematics Mompati S. Koorapetse P. Kaelo Globally convergent three-term conjugate gradient projection methods
More informationConjugate gradient methods based on secant conditions that generate descent search directions for unconstrained optimization
Conjugate gradient methods based on secant conditions that generate descent search directions for unconstrained optimization Yasushi Narushima and Hiroshi Yabe September 28, 2011 Abstract Conjugate gradient
More informationTHE RELATIONSHIPS BETWEEN CG, BFGS, AND TWO LIMITED-MEMORY ALGORITHMS
Furman University Electronic Journal of Undergraduate Mathematics Volume 12, 5 20, 2007 HE RELAIONSHIPS BEWEEN CG, BFGS, AND WO LIMIED-MEMORY ALGORIHMS ZHIWEI (ONY) QIN Abstract. For the solution of linear
More informationGlobal Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method. 1 Introduction
ISSN 1749-3889 (print), 1749-3897 (online) International Journal of Nonlinear Science Vol.11(2011) No.2,pp.153-158 Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method Yigui Ou, Jun Zhang
More information230 L. HEI if ρ k is satisfactory enough, and to reduce it by a constant fraction (say, ahalf): k+1 = fi 2 k (0 <fi 2 < 1); (1.7) in the case ρ k is n
Journal of Computational Mathematics, Vol.21, No.2, 2003, 229 236. A SELF-ADAPTIVE TRUST REGION ALGORITHM Λ1) Long Hei y (Institute of Computational Mathematics and Scientific/Engineering Computing, Academy
More informationMath 164: Optimization Barzilai-Borwein Method
Math 164: Optimization Barzilai-Borwein Method Instructor: Wotao Yin Department of Mathematics, UCLA Spring 2015 online discussions on piazza.com Main features of the Barzilai-Borwein (BB) method The BB
More informationHT14 A COMPARISON OF DIFFERENT PARAMETER ESTIMATION TECHNIQUES FOR THE IDENTIFICATION OF THERMAL CONDUCTIVITY COMPONENTS OF ORTHOTROPIC SOLIDS
Inverse Problems in Engineering: heory and Practice rd Int. Conference on Inverse Problems in Engineering June -8, 999, Port Ludlow, WA, USA H4 A COMPARISON OF DIFFEREN PARAMEER ESIMAION ECHNIQUES FOR
More informationNew Inexact Line Search Method for Unconstrained Optimization 1,2
journal of optimization theory and applications: Vol. 127, No. 2, pp. 425 446, November 2005 ( 2005) DOI: 10.1007/s10957-005-6553-6 New Inexact Line Search Method for Unconstrained Optimization 1,2 Z.
More informationQuasi-Newton Methods
Quasi-Newton Methods Werner C. Rheinboldt These are excerpts of material relating to the boos [OR00 and [Rhe98 and of write-ups prepared for courses held at the University of Pittsburgh. Some further references
More informationDownloaded 12/02/13 to Redistribution subject to SIAM license or copyright; see
SIAM J. OPTIM. Vol. 23, No. 4, pp. 2150 2168 c 2013 Society for Industrial and Applied Mathematics THE LIMITED MEMORY CONJUGATE GRADIENT METHOD WILLIAM W. HAGER AND HONGCHAO ZHANG Abstract. In theory,
More information1 Numerical optimization
Contents 1 Numerical optimization 5 1.1 Optimization of single-variable functions............ 5 1.1.1 Golden Section Search................... 6 1.1. Fibonacci Search...................... 8 1. Algorithms
More informationLine Search Methods for Unconstrained Optimisation
Line Search Methods for Unconstrained Optimisation Lecture 8, Numerical Linear Algebra and Optimisation Oxford University Computing Laboratory, MT 2007 Dr Raphael Hauser (hauser@comlab.ox.ac.uk) The Generic
More informationGradient method based on epsilon algorithm for large-scale nonlinearoptimization
ISSN 1746-7233, England, UK World Journal of Modelling and Simulation Vol. 4 (2008) No. 1, pp. 64-68 Gradient method based on epsilon algorithm for large-scale nonlinearoptimization Jianliang Li, Lian
More information1 Numerical optimization
Contents Numerical optimization 5. Optimization of single-variable functions.............................. 5.. Golden Section Search..................................... 6.. Fibonacci Search........................................
More informationA new initial search direction for nonlinear conjugate gradient method
International Journal of Mathematis Researh. ISSN 0976-5840 Volume 6, Number 2 (2014), pp. 183 190 International Researh Publiation House http://www.irphouse.om A new initial searh diretion for nonlinear
More informationGRADIENT METHODS FOR LARGE-SCALE NONLINEAR OPTIMIZATION
GRADIENT METHODS FOR LARGE-SCALE NONLINEAR OPTIMIZATION By HONGCHAO ZHANG A DISSERTATION PRESENTED TO THE GRADUATE SCHOOL OF THE UNIVERSITY OF FLORIDA IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE
More informationNONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM
NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM JIAYI GUO AND A.S. LEWIS Abstract. The popular BFGS quasi-newton minimization algorithm under reasonable conditions converges globally on smooth
More informationUnconstrained optimization
Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout
More informationA line-search algorithm inspired by the adaptive cubic regularization framework, with a worst-case complexity O(ɛ 3/2 ).
A line-search algorithm inspired by the adaptive cubic regularization framewor, with a worst-case complexity Oɛ 3/. E. Bergou Y. Diouane S. Gratton June 16, 017 Abstract Adaptive regularized framewor using
More informationGradient methods exploiting spectral properties
Noname manuscript No. (will be inserted by the editor) Gradient methods exploiting spectral properties Yaui Huang Yu-Hong Dai Xin-Wei Liu Received: date / Accepted: date Abstract A new stepsize is derived
More informationStatistics 580 Optimization Methods
Statistics 580 Optimization Methods Introduction Let fx be a given real-valued function on R p. The general optimization problem is to find an x ɛ R p at which fx attain a maximum or a minimum. It is of
More informationA class of Smoothing Method for Linear Second-Order Cone Programming
Columbia International Publishing Journal of Advanced Computing (13) 1: 9-4 doi:1776/jac1313 Research Article A class of Smoothing Method for Linear Second-Order Cone Programming Zhuqing Gui *, Zhibin
More informationA DIMENSION REDUCING CONIC METHOD FOR UNCONSTRAINED OPTIMIZATION
1 A DIMENSION REDUCING CONIC METHOD FOR UNCONSTRAINED OPTIMIZATION G E MANOUSSAKIS, T N GRAPSA and C A BOTSARIS Department of Mathematics, University of Patras, GR 26110 Patras, Greece e-mail :gemini@mathupatrasgr,
More informationA Conjugate Gradient Method with Inexact. Line Search for Unconstrained Optimization
Applie Mathematical Sciences, Vol. 9, 5, no. 37, 83-83 HIKARI Lt, www.m-hiari.com http://x.oi.or/.988/ams.5.4995 A Conuate Graient Metho with Inexact Line Search for Unconstraine Optimization * Mohame
More informationSOLVING SYSTEMS OF NONLINEAR EQUATIONS USING A GLOBALLY CONVERGENT OPTIMIZATION ALGORITHM
Transaction on Evolutionary Algorithm and Nonlinear Optimization ISSN: 2229-87 Online Publication, June 2012 www.pcoglobal.com/gjto.htm NG-O21/GJTO SOLVING SYSTEMS OF NONLINEAR EQUATIONS USING A GLOBALLY
More informationON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS
ON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS Anders FORSGREN Tove ODLAND Technical Report TRITA-MAT-203-OS-03 Department of Mathematics KTH Royal
More informationChapter 4. Unconstrained optimization
Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file
More informationDENSE INITIALIZATIONS FOR LIMITED-MEMORY QUASI-NEWTON METHODS
DENSE INITIALIZATIONS FOR LIMITED-MEMORY QUASI-NEWTON METHODS by Johannes Brust, Oleg Burdaov, Jennifer B. Erway, and Roummel F. Marcia Technical Report 07-, Department of Mathematics and Statistics, Wae
More informationA new class of Conjugate Gradient Methods with extended Nonmonotone Line Search
Appl. Math. Inf. Sci. 6, No. 1S, 147S-154S (2012) 147 Applied Mathematics & Information Sciences An International Journal A new class of Conjugate Gradient Methods with extended Nonmonotone Line Search
More informationA PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES
IJMMS 25:6 2001) 397 409 PII. S0161171201002290 http://ijmms.hindawi.com Hindawi Publishing Corp. A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES
More informationModification of the Wolfe line search rules to satisfy the descent condition in the Polak-Ribière-Polyak conjugate gradient method
Laboratoire d Arithmétique, Calcul formel et d Optimisation UMR CNRS 6090 Modification of the Wolfe line search rules to satisfy the descent condition in the Polak-Ribière-Polyak conjugate gradient method
More informationResearch Article Finding Global Minima with a Filled Function Approach for Non-Smooth Global Optimization
Hindawi Publishing Corporation Discrete Dynamics in Nature and Society Volume 00, Article ID 843609, 0 pages doi:0.55/00/843609 Research Article Finding Global Minima with a Filled Function Approach for
More informationHigher-Order Methods
Higher-Order Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. PCMI, July 2016 Stephen Wright (UW-Madison) Higher-Order Methods PCMI, July 2016 1 / 25 Smooth
More informationNonmonotonic back-tracking trust region interior point algorithm for linear constrained optimization
Journal of Computational and Applied Mathematics 155 (2003) 285 305 www.elsevier.com/locate/cam Nonmonotonic bac-tracing trust region interior point algorithm for linear constrained optimization Detong
More informationModification of Non-Linear Constrained Optimization. Ban A. Metras Dep. of Statistics College of Computer Sciences and Mathematics
Raf. J. Of Comp. & Math s., Vol., No., 004 Modification of Non-Linear Constrained Optimization Ban A. Metras Dep. of Statistics College of Computer Sciences and Mathematics Received: 3/5/00 Accepted: /0/00
More informationNonlinear Programming
Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week
More informationFALL 2018 MATH 4211/6211 Optimization Homework 4
FALL 2018 MATH 4211/6211 Optimization Homework 4 This homework assignment is open to textbook, reference books, slides, and online resources, excluding any direct solution to the problem (such as solution
More informationConvergence rate of SGD
Convergence rate of SGD heorem: (see Nemirovski et al 09 from readings) Let f be a strongly convex stochastic function Assume gradient of f is Lipschitz continuous and bounded hen, for step sizes: he expected
More informationModification of the Wolfe line search rules to satisfy the descent condition in the Polak-Ribière-Polyak conjugate gradient method
Laboratoire d Arithmétique, Calcul formel et d Optimisation UMR CNRS 6090 Modification of the Wolfe line search rules to satisfy the descent condition in the Polak-Ribière-Polyak conjugate gradient method
More informationAcceleration Method for Convex Optimization over the Fixed Point Set of a Nonexpansive Mapping
Noname manuscript No. will be inserted by the editor) Acceleration Method for Convex Optimization over the Fixed Point Set of a Nonexpansive Mapping Hideaki Iiduka Received: date / Accepted: date Abstract
More informationUnconstrained Multivariate Optimization
Unconstrained Multivariate Optimization Multivariate optimization means optimization of a scalar function of a several variables: and has the general form: y = () min ( ) where () is a nonlinear scalar-valued
More informationOn the acceleration of augmented Lagrangian method for linearly constrained optimization
On the acceleration of augmented Lagrangian method for linearly constrained optimization Bingsheng He and Xiaoming Yuan October, 2 Abstract. The classical augmented Lagrangian method (ALM plays a fundamental
More information17 Solution of Nonlinear Systems
17 Solution of Nonlinear Systems We now discuss the solution of systems of nonlinear equations. An important ingredient will be the multivariate Taylor theorem. Theorem 17.1 Let D = {x 1, x 2,..., x m
More informationJournal of Computational and Applied Mathematics. Notes on the Dai Yuan Yuan modified spectral gradient method
Journal of Computational Applied Mathematics 234 (200) 2986 2992 Contents lists available at ScienceDirect Journal of Computational Applied Mathematics journal homepage: wwwelseviercom/locate/cam Notes
More informationmin f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term;
Chapter 2 Gradient Methods The gradient method forms the foundation of all of the schemes studied in this book. We will provide several complementary perspectives on this algorithm that highlight the many
More informationMultipoint secant and interpolation methods with nonmonotone line search for solving systems of nonlinear equations
Multipoint secant and interpolation methods with nonmonotone line search for solving systems of nonlinear equations Oleg Burdakov a,, Ahmad Kamandi b a Department of Mathematics, Linköping University,
More informationResidual iterative schemes for largescale linear systems
Universidad Central de Venezuela Facultad de Ciencias Escuela de Computación Lecturas en Ciencias de la Computación ISSN 1316-6239 Residual iterative schemes for largescale linear systems William La Cruz
More informationA Novel of Step Size Selection Procedures. for Steepest Descent Method
Applied Mathematical Sciences, Vol. 6, 0, no. 5, 507 58 A Novel of Step Size Selection Procedures for Steepest Descent Method Goh Khang Wen, Mustafa Mamat, Ismail bin Mohd, 3 Yosza Dasril Department of
More informationA line-search algorithm inspired by the adaptive cubic regularization framework with a worst-case complexity O(ɛ 3/2 )
A line-search algorithm inspired by the adaptive cubic regularization framewor with a worst-case complexity Oɛ 3/ E. Bergou Y. Diouane S. Gratton December 4, 017 Abstract Adaptive regularized framewor
More informationSolutions and Notes to Selected Problems In: Numerical Optimzation by Jorge Nocedal and Stephen J. Wright.
Solutions and Notes to Selected Problems In: Numerical Optimzation by Jorge Nocedal and Stephen J. Wright. John L. Weatherwax July 7, 2010 wax@alum.mit.edu 1 Chapter 5 (Conjugate Gradient Methods) Notes
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428
More informationA Trust Region Algorithm Model With Radius Bounded Below for Minimization of Locally Lipschitzian Functions
The First International Symposium on Optimization and Systems Biology (OSB 07) Beijing, China, August 8 10, 2007 Copyright 2007 ORSC & APORC pp. 405 411 A Trust Region Algorithm Model With Radius Bounded
More informationIntroduction to Nonlinear Optimization Paul J. Atzberger
Introduction to Nonlinear Optimization Paul J. Atzberger Comments should be sent to: atzberg@math.ucsb.edu Introduction We shall discuss in these notes a brief introduction to nonlinear optimization concepts,
More informationOn Lagrange multipliers of trust-region subproblems
On Lagrange multipliers of trust-region subproblems Ladislav Lukšan, Ctirad Matonoha, Jan Vlček Institute of Computer Science AS CR, Prague Programy a algoritmy numerické matematiky 14 1.- 6. června 2008
More informationA NOVEL FILLED FUNCTION METHOD FOR GLOBAL OPTIMIZATION. 1. Introduction Consider the following unconstrained programming problem:
J. Korean Math. Soc. 47, No. 6, pp. 53 67 DOI.434/JKMS..47.6.53 A NOVEL FILLED FUNCTION METHOD FOR GLOBAL OPTIMIZATION Youjiang Lin, Yongjian Yang, and Liansheng Zhang Abstract. This paper considers the
More informationOPER 627: Nonlinear Optimization Lecture 14: Mid-term Review
OPER 627: Nonlinear Optimization Lecture 14: Mid-term Review Department of Statistical Sciences and Operations Research Virginia Commonwealth University Oct 16, 2013 (Lecture 14) Nonlinear Optimization
More informationJae Heon Yun and Yu Du Han
Bull. Korean Math. Soc. 39 (2002), No. 3, pp. 495 509 MODIFIED INCOMPLETE CHOLESKY FACTORIZATION PRECONDITIONERS FOR A SYMMETRIC POSITIVE DEFINITE MATRIX Jae Heon Yun and Yu Du Han Abstract. We propose
More informationNumerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09
Numerical Optimization 1 Working Horse in Computer Vision Variational Methods Shape Analysis Machine Learning Markov Random Fields Geometry Common denominator: optimization problems 2 Overview of Methods
More informationSteepest descent algorithm. Conjugate gradient training algorithm. Steepest descent algorithm. Remember previous examples
Conjugate gradient training algorithm Steepest descent algorithm So far: Heuristic improvements to gradient descent (momentum Steepest descent training algorithm Can we do better? Definitions: weight vector
More informationGlobal convergence of a regularized factorized quasi-newton method for nonlinear least squares problems
Volume 29, N. 2, pp. 195 214, 2010 Copyright 2010 SBMAC ISSN 0101-8205 www.scielo.br/cam Global convergence of a regularized factorized quasi-newton method for nonlinear least squares problems WEIJUN ZHOU
More information