Nonlinear Programming
|
|
- Mervin Bryan
- 5 years ago
- Views:
Transcription
1 Nonlinear Programming Kees Roos URL: roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D Optimization Group 1
2 Outline for week 7: Algorithms for unconstrained minimization A generic algorithm Rate of convergence Line search methods Dichotomous and golden section search Bisection Newton s method Search directions Gradient method Newton s method Methods of conjugate directions Powell s method Fletcher-Reeves method Quasi-Newton s method DFP update BFGS update Stopping criteria Optimization Group 2
3 Input: ǫ > 0 is the accuracy parameter; x 0 is a given (relative interior) feasible point; Generic algorithm for min x C f(x) Step 0: x := x 0, k = 0; Step 1: Find search direction s k s.t. δf(x k, s k ) < 0 (This should be a descending feasible direction in the constrained case.) Step 1a: If no such direction exists STOP, optimum found. Step 2: Line search : find λ k = argmin λ f(x k + λs k ); Step 3: x k+1 = x k + λ k s k, k = k + 1; Step 4: If stopping criteria satisfied STOP, else GOTO Step 1. Optimization Group 3
4 Algorithms: rate of convergence Definition: Let α 1, α 2,..., α k,... α be a convergent sequence. The rate of convergence is: p = sup { p : limsup k The larger p is, the faster the convergence. Let The rate of convergence is: β = limsup k α k+1 α α k α p < α k+1 α α k α p. }. linear if p = 1 and 0 < β < 1; super-linear if p = 1 and β = 0; quadratic if p = 2; sub-linear if β = 1. Optimization Group 4
5 Examples: order of convergence Example 1: The sequence α k = a k, where 0 < a < 1 converges linearly to zero while β = a. Example 2: The sequence α k = a (2k), where 0 < a < 1 converges quadratically to zero. Example 3: The sequence α k = 1 k converges sub-linearly to zero. Example 4: The sequence α k = ( 1 k )k converges super-linearly to zero. Optimization Group 5
6 Linesearch methods We assume throughout that f is a convex function. We are given a (feasible) search direction s at a feasible point x and we want to find So we are minimizing for λ 0. λ = argmin λ 0 f(x + λs). φ(λ) := f(x + λs) This is a one-dimensional problem. We deal with four different line search methods, that require different levels of information about φ(λ): The Dichotomous search and Golden section methods, that use only function evaluations of φ; Bisection, that evaluates φ (λ) (φ has to be continuously differentiable); Newton s method, that evaluates both φ (λ) and φ (λ). Optimization Group 6
7 Linesearch: Dichotomous search We assume that φ is convex and has a minimizer on the interval [a, b]. Our aim is to reduce the size of this interval of uncertainty by evaluating φ at points in [a, b]. Lemma 1 (Exercise 4.7) Let a < ā < b < b. If φ(ā) < φ( b) then the minimum of φ occurs in the interval [a, b]; if φ(ā) φ( b) the minimum of φ occurs in the interval [ā, b]. The lemma suggest a simple algorithm to reduce the interval of uncertainty. Optimization Group 7
8 Input: ǫ > 0 is the accuracy parameter; Linesearch: Dichotomous search a 0, b 0 are given such that [a 0, b 0 ] contains the minimizer of φ(λ), k = 0. Step 1: If a k b k < ǫ STOP. Step 2: Choose ā k (a k, b k ) and b k (a k, b k ), such that ā k < b k ; Step 3a: If φ(ā k ) < φ( b k ) set a k+1 = a k, b k+1 = b k ; Step 3b: If φ(ā k ) φ( b k ) set a k+1 = ā k, b k+1 = b k ; Step 4: Set k = k + 1. GOTO Step 1. We have not specified yet how we should choose the values ā k and b k in iteration k (Step 2 of the algorithm). There are many ways to do this. One is to choose ā k = 1 2 (a k + b k ) δ and b k = 1 2 (a k + b k ) + δ where δ > 0 is a (very) small fixed constant. Then the interval of uncertainty is reduced by a factor ( δ)t/2 after t function evaluations (Exercise 4.8). Optimization Group 8
9 Linesearch: Golden section search This is a variant of the Dichotomous search method where the constant δ is not constant but depends on k. In the k-th iteration we take δ = δ k, where δ k = ( α 1 2 ) (b k a k ), α = 1 2 ( 5 1 ) Here α is the Golden ratio, i.e., the root of α 2 + α 1 = 0 with α [0,1]. We now have ā k = 1 2 (a k + b k ) δ k = 1 2 (a k + b k ) ( α 1 2 (b k a k ) = b k α(b k a k ) b k = 1 2 (a k + b k ) + δ k = 1 2 (a k + b k ) + ( α 1 ) 2 (b k a k ) = a k + α(b k a k ). If φ(ā k ) < φ( b k ), then we set b k+1 = b k and a k+1 = a k. In that case b k+1 = a k+1 + α ( b k+1 a k+1 ) = ak + α ( b k a k ) = ak + α 2 (b k a k ) = a k + (1 α)(b k a k ) = b k α(b k a k ) = ā k. So in the next iteration we only need to compute φ(ā k+1 ). Similarly, if φ(ā k ) φ( b k ), then we set a k+1 = ā k and b k+1 = b k, and it follows in a similar way that ā k+1 = b k. So in the next iteration we only need to compute φ( b k+1 ). In both cases one need to evaluate φ only once. See the course notes for graphical illustrations. When using Golden section search each iteration reduces the interval of uncertainty by a factor α (Exercise 4.9). ) Optimization Group 9
10 Linesearch: Golden section search Suppose a k = 0 and b k = 1. We choose fixed α ( 1 2,1) and define ā k+1 = a k + (1 α)(b k a k ) = 1 α, b k+1 = b k (1 α)(b k a k ) = α. Suppose φ(1 α) < φ(α). Then we set a k+1 = 0 and b k+1 = α and ā k+2 = a k+1 + (1 α)(b k+1 a k+1 ) = (1 α)α b k+2 = b k+1 (1 α)(b k+1 a k+1 ) = α (1 α)α. We want one of these two points to be ā k+1 = 1 α, because we already know φ(1 α). This gives either or, equivalently, (1 α)α = 1 α or α α(1 α) = 1 α, α = 1 or α 2 + α 1 = 0. Since α ( 1 2,1), the only possible value for α is which is the Golden ratio! α = 1 2 ( 5 1 ) 0.618, Optimization Group 10
11 Linesearch: Bisection (or Bolzano s method) We assume that φ(λ) is differentiable (and convex). We wish to find a λ so that φ ( λ) = 0. Input: ǫ > 0 is the accuracy parameter; a 0, b 0 are given such that φ (a 0 ) < 0 & φ (b 0 ) > 0, k = 0; Step 1: If b k a k < ǫ STOP. Step 2: Let λ = 1 2 (a k + b k ); Step 3a: If φ (λ) < 0 then set a k+1 = λ, b k+1 = b k ; Step 3b: If φ (λ) > 0 then set a k+1 = a k, b k+1 = λ; Step 4: Set k = k + 1. GOTO Step 1. The algorithm needs log 2 b 0 a 0 ǫ evaluations of φ (Exercise 4.11). Optimization Group 11
12 Linesearch using the Newton-Raphson method The quadratic approximation of φ at λ k is q(λ) = φ(λ k ) + φ (λ k )(λ λ k ) φ (λ k )(λ λ k ) 2. The minimum of q is attained if q (λ) = 0, which gives Input: ǫ > 0 is the accuracy parameter; λ 0 is the given initial point; k = 0; λ k+1 = λ k φ (λ k ) φ (λ k ). Step 1: Let λ k+1 = λ k φ (λ k ) φ (λ k ) ; Step 2: If λ k+1 λ k < ǫ STOP. Step 3: k := k + 1, GOTO Step 1. Optimization Group 12
13 The Newton-Raphson method: Example 4.3 Let φ(λ) = λ log(1+λ). The domain of φ is ( 1, ). The first and second derivatives of φ are φ (λ) = λ 1 + λ, φ (λ) = 1 (1 + λ) 2. This makes clear that φ is strictly convex on its domain, and minimal at λ = 0. The iterates satisfy the recursive relation λ k+1 = λ k φ (λ k ) φ (λ k ) = λ k λ k (1 + λ k ) = λ 2 k. This implies quadratic convergence if λ 0 < 1 (see Exercise 4.12). On the other hand, Newton s method fails if λ 0 1. For example, if λ 0 = 1 then λ 1 = 1, which is not in the domain of φ! In general the method converges quadratically if the following conditions are met: 1. the starting point is sufficiently close to the minimizer; 2. in addition to being convex, the function φ has a property called self-concordance, which is introduced later. Optimization Group 13
14 Search Directions: The Gradient method Search direction: s = f(x k ) Steepest descent direction! δf(x, f(x)) = f(x) T f(x) = min s = f(x) { f(x) T s}. The (negative) gradient is orthogonal to the level curves (Exercise 4.14). The gradient method is not a finite algorithm, not even for linear or quadratic functions. Slow convergence ( zigzagging ) (Figure 4.4). The order of convergence is only linear. Optimization Group 14
15 Let f be continuously differentiable. Convergence of the gradient method Start from the initial point x 0 using exact line search; the gradient method produces a sequence { x k} such that f(x k ) > f(x k+1 ) for k = 0,1,2,. Assume that the level set D = {x : f(x) f(x 0 )} is compact, then any accumulation point x of the sequence { x k} is a stationary point (i.e. f( x) = 0) of f. If the function f is a convex function, then x is a global minimizer of f. If the function f is not convex, then x is a local minimizer of f. Optimization Group 15
16 Newton s method Newton s method is based on minimizing the second order approximation of f at x k. q(x) := f(x k ) + f(x k ) T (x x k ) (x xk ) T 2 f(x k )(x x k ). We assume that q(x) is strictly convex. So the Hessian 2 f(x k ) is positive definite. Hence the minimum is attained when q(x) = f(x k ) + 2 f(x k )(x x k ) = 0. We can solve x from 2 f(x k )(x x k ) = f(x k ) which gives the next iterate: x k+1 = x k ( 2 f(x k )) 1 f(x k ), So the Newton direction is s k = ( 2 f(x k )) 1 f(x k ). Exact when f is quadratic. Local quadratic convergence with full Newton steps (α = 1, so without any line search!). Good starting point is essential. Optimization Group 16
17 Trust region method If the function f(x) is not strictly convex, or if the Hessian is ill-conditioned, then the Hessian is not (or hardly) invertible. Remedy: trust region method. 2 f(x) is replaced by ( 2 f(x) + αi); s k = ( 2 f(x k ) + αi) 1 f(x k ). α is dynamically increased and decreased in order to avoid exact line search. If α = 0 then we have the Newton step, if α then we approach a (small) multiple of the negative gradient. Optimization Group 17
18 Newton s method for solving nonlinear equations Find solution of F(x) = 0, where F : R m R n Linearize at x k : F(x) F(x k ) + JF(x k )(x x k ) Jacobian of F: JF(x) ij = F i(x) x j. Solve x k+1 from JF(x k )(x k+1 x k ) = F(x k ). Minimizing f(x) is equivalent to solving f(x) = 0. 2 f(x k )(x k+1 x k ) = f(x k ). The Jacobian of the gradient is exactly the Hessian of the function f(x) hence it is positive definite and we have as we have seen above. x k+1 = x k ( 2 f(x k )) 1 f(x k ) Conclusion: Newton s optimization method Newton s method for nonlinear equations applied to the system f(x) = 0. Optimization Group 18
19 Methods using conjugate directions (1) Let A be an n n symmetric positive definite matrix and b R n. We consider min { q(x) = 1 2 xt Ax b T x : x R n}. The minimizer is uniquely determined by q(x) = Ax b = 0. But to find the minimizer we need to invert the matrix A. If n is large this is computationally expensive, and we want to avoid this. This can be done by using so-called conjugate search directions. If the subsequent search directions are s 0,..., s k then the iterates have the form x k+1 = x k + λ k s k, k = 1, 2,.... If we use exact line search, then we automatically have q(x k+1 ) T s k = 0, k = 1, 2,.... By requiring a little more, namely that the search vectors s i are linearly independent and q(x k+1 ) T s i = 0, 0 i k = 1, 2,... we can guarantee termination of the algorithm in a finite number of steps. Because then q(x n ) T s i = 0, i < n, whence, since the vectors s i are linearly independent, q(x n ) = 0. So no more than n steps are required. Optimization Group 19
20 Methods using conjugate directions (2) We denote q(x k ) as g k. Note that when using exact line search, we have (automatically) g j+1t s j = 0, j = 1, 2,.... Lemma 2 Let k {1,..., n}. The following two statement are equivalent: (i) g j+1t s i = 0, 0 i < j k; (ii) s it As j = 0, 0 i < j k. Proof: Since q(x) = Ax b, we have q(x j+1 ) = Ax j+1 b = A ( x j + λ j s j) b = q(x j ) + λ j As j = g j + λ j As j, j = 0, 1,... Therefore, for each i 0, g j+1t s i = q(x j+1 ) T s i = g jt s i + λ j s it As j, j = 0, 1,.... The proof can now be easily completed by induction on k, since λ j > 0 for each j. If (ii) holds then the vectors s 0,..., s k R n are called conjugate (or A conjugate). Note that if A = I then conjugate means orthogonal. Then s 0,..., s k are linearly independent. This also holds for A-conjugate vectors, since A is positive definite (Ex. 4.20). As we established before, if one uses A-conjugate directions to minimize the quadratic form q, then the minimizer of q is found in at most n iterations. Optimization Group 20
21 Easy method to generate conjugate directions Let s 0 = q(x 0 ) = g 0. Then we can get subsequent conjugate directions by taking s k = g k + α k s k 1, k = 1, 2,... for suitable values of α k. In order to make s k and s k 1 A-conjugate, we must have s kt As k 1 = 0 for k 1. This determines the coefficients α k uniquely: α k = gkt As k 1 k 1. s k 1T Ask 1, We proceed with induction on k. So we assume that s 0,..., s k 1 are conjugate. Using g kt s k 1 = 0 we find g kt s k = g kt ( g k + α k s k 1) = g k 2 < 0, proving that s k is a descent direction, provided g k 0. Our choice of α k implies s kt As k 1 = 0. So it remains to show that s kt As i = 0 for i < k 1. The induction hypothesis implies s kt As i = ( g k + α k s k 1) T As i = g kt As i. Since g i = q(x i ) = Ax i b, and x i+1 = x i + λ i s i, we have Hence, due to Lemma 2(ii), λ i As i = g i+1 g i = α i+1 s i s i+1 ( α i s i 1 s i). λ i g kt As i = g kt ( αi+1 s i s i+1 α i s i 1 + s i) = 0. This proves that s 0,..., s k are conjugate, provided g k 0 (otherwise x k is optimal!). Optimization Group 21
22 The case of nonquadratic functions In the case where f is (convex) quadratic then finite termination is guaranteed if ( ) ( ) α k = gkt As k 1 s k 1T As = gkt gk gk 1 k 1 s k 1T (g k g k 1 ) = gkt gk gk 1 g k 2 g k 1 2 = g k 1 2, k 1. Here we used that g k g k 1, as follows from g k 1 = s k 1 +α k s k 2 and g kt s k 2 = 0, by Lemma 2 (i) and g kt s k 1 = 0 by the choice of λ k. The algorithm is Step 0. Let s 0 = f(x 0 ) and x 1 := argmin f(x 0 + λs 0 ). Step k. Set s k = f(x k ) + α k s k 1 and x k+1 := argmin f(x k + λs k ). If f is not quadratic there is no guarantee that the method stops after a finite number of steps. Several choices for α k have been proposed (which are equivalent in the quadratic case): ( Hestenes-Stiefel (1952): α k = gkt g k g k 1) s k 1T (g k g k 1 ). g k 2 Fletcher-Reeves (1964): α k = g k 1 2. Polak-Ribière (1969): α k = gkt ( g k g k 1) g k 1 2. Optimization Group 22
23 Solving a linear system with the conjugate gradient method Assume we want to solve Ax = b with A positive definite. The solution is precisely the minimizer of q(x) = 1 2 xt Ax b T x, and hence can be solve by the conjugate gradient method. If A is not positive definite, but nonsingular, then A T A is positive definite. Hence we can solve Ax = b by minimizing q(x) = Ax b 2 = x T A T Ax 2b T Ax + b T b. Optimization Group 23
24 Powell s method We now deal with a conjugate direction method using only function values (no gradients!). Input A starting point x 0, a set of linearly independent vectors t 1,..., t n. Initialization Set t (1,i) = t i, i = 1,..., n. For k = 1,..., n do (Cycle k) Let z (k,1) = x k 1 and z (k,i+1) := argmin q ( z (k,i) + λt (k,i)), i = 1,..., n. Let x k := argmin q(z (k,n+1) + λs k ), where s k := z (k,n+1) x k 1. Let t (k+1,i) = t (k,i+1), i = 1,, n 1 and t (k+1,n) := s k. The algorithm consists of n cycles and terminates at the minimizer of q(x). Each cycle consists of n + 1 line searches and yields a search direction s k. The k-th direction s k is constructed at the end of cycle k. The search directions s 1,..., s n are conjugate (for a proof: see the course notes). Note that only function values are evaluated (no derivatives are used, unless the line searches do so). The number of line searches is n(n + 1). Therefore, Powell s method is attractive for minimizing black box functions where gradient and Hessian information is not available (or too expensive to compute). Optimization Group 24
25 Illustration of Powell s method 2 x t x 1 x 2 0 x 2 optimal t x 1 Iterates of Powell s algorithm for f(x) = 5x x 1x 2 + x , starting at x0 = [1;2]. Optimization Group 25
26 Quasi-Newton methods Recall that the Newton direction at iteration k is given by: s k = [ 2 f(x k ) ] 1 f(x k ) = [ 2 f(x k ) ] 1 g k. Quasi-Newton methods use a positive definite approximation H k to [ 2 f(x k ) ] 1. The approximation H k is updated at each iteration, say H k+1 = H k + D k, where D k denotes the update. The algorithm has the following generic form. Step 0. Let x 0 be given and set H 0 = I. Step k. s k = H k g k and x k+1 = argmin λ f(x k + λs k ) = x k + λ k s k ; H k+1 = H k + D k and k = k + 1. Defining we require for each k that y k := g k+1 g k, σ k = x k+1 x k = λ k s k. I H k+1 is symmetric positive definite; II σ k = H k+1 y k (quasi-newton property); III σ i = H k+1 y i (i = 0,..., k 1) (hereditary property), Optimization Group 26
27 The quasi-newton property and hereditary property Let A be an n n symmetric PD matrix, and let f be the strictly convex quadratic function Then g k = q(x k ) = Ax k b, and hence q(x) = 1 2 xt Ax b T x. y k = g k+1 g k = q(x k+1 ) q(x k ) = A ( x k+1 x k) = Aσ k, whence σ k = A 1 y k. Recall that each H k should be a good approximation of the inverse of 2 q(x k ), which is A 1. Therefore we require that σ k = H k+1 y k, which is the quasi-newton property, and even more, that our approximation H k satisfies σ i = H k+1 y i, i = 0,..., k. which is the hereditary property. Note that the hereditary property implies σ i = H n y i, i = 0,..., n 1. If the σ i (i = 0,..., n 1) are linearly independent, this implies H n = A 1. But then the n + 1-th iteration is simply the Newton step at x n. Since q is quadratic, this yields the minimizer of q, and hence we find the minimum of q no more than n + 1 iterations. Optimization Group 27
28 A generic update D k (1) First consider the case where D k is a (possibly indefinite) matrix of rank 2, whence D k = αuu T + βvv T for suitable vectors u and v and scalars α, β. Then the quasi-newton property implies H k+1 y k = H k y k + αuu T y k + βvv T y k = σ k. Davidon, Fletcher and Powell (1963) recognized that this condition is satisfied if u = σ k = λ k s k, α = 1, v = H k y k 1, β = σ kt y k y kt H k y k, which yields the so-called DFP update: D k = λ k s k s kt s kt y k H kyk ykt Hk y kt H k y k. In the following we consider a slight more general update, namely D k = αuu T + βvv T + µ ( uv T + vu T ) = [ ] u v α µ µ β Exercise A Show that D k has rank at most 2. ut v T. Optimization Group 28
29 A generic update D k (2) D k = αuu T + βvv T + µ ( uv T + vu T) = [ u v ] [ α µ µ β ] [ u T v T ], u = σ k, v = H k y k. Lemma 3 If the above update D k satisfies the quasi-newton property then the subsequent directions are conjugate. (So a quasi-newton method is a conjugate gradient method!) Proof: We show by induction on k that H k y i = σ i = λ i s i, s kt As i = g kt s i = 0, 0 i < k. (1) This trivially holds if k = 0 (the condition is void). Assuming the quasi-newton property and (1) for k 0, and using y i = Aσ i and σ i = λ i s i for all i, we write for i < k: y kt H k+1 y i = σ kt y i = σ kt Aσ i = λ k λ i s kt As i = 0. Also σ T k y i = σ T k Aσ i = 0. Hence we obtain for all i < k, D k y i = [ u v ] [ ] [ ] α µ u T y i µ β v T = [ σ k H k y ] [ k α µ y i µ β ] [ σ kt y i y kt H k y i whence H k+1 y i = H k y i +D k y i = H k y i = σ i. Together with the quasi-newton property this gives H k+1 y i = σ i for 0 i < k + 1. Because λ i 0, s k+1 = H k+1 g k+1 and H k+1 y i = σ i we observe next that λ i s k+1t As i = s k+1t Aσ i = s k+1t y i = g k+1t H k+1 y i = g k+1t σ i. Hence it suffices for the rest of the proof if g k+1t s i = 0 for 0 i < k + 1. This certainly holds if i = k, because we use exact line search. For i < k we use the induction hypothesis again, and g k+1 = A(x k + λ k s k ) b = g k + λ k As k, which gives This completes the proof. g k+1t s i = g kt s i + λ k s kt As = 0. ] = 0, Optimization Group 29
30 The Broyden family of updates (1) D k = αuu T + βvv T + µ ( uv T + vu T) = [ u v ] [ α µ µ β ] [ u T v T ], u = σ k, v = H k y k. We now determine conditions on the parameters α, β and µ that guarantee that each H k is positive definite. The quasi-newton property (σ k = H k+1 y k = H k y k + D k y k for k 0) amounts to u = v + αuu T y k + βvv T y k + µ ( uv T + vu T ) y k To satisfy this condition it suffices if = v ( 1 + βv T y k + µu T y k) + u ( αu T y k + µv T y k). αu T y k + µv T y k = 1 βv T y k + µu T y k = 1. This linear system has multiple solutions. Introducing ρ = µu T y k R the solution is α = 1 u T y k Since v T y k = y T H k y k > 0 and ( 1 + ρ vt y k u T y k ), β = ρ 1 v T y k, ρ R. u T y k = σ k T ( g k+1 g k) = λ k s k T ( g k+1 g k) = λ k s k T g k = λ k g k T H k g k > 0, the above expressions are well defined. Optimization Group 30
31 The Broyden family of updates (2) D k = αuu T +βvv T +µ ( uv T + vu T), u = σ k, v = H k y k, α = 1 u T y k Substituting the values of u, v, α and β we find D k = λ k s k s kt s kt y k H kyk ykt Hk y kt H k y k +ρwwt, w = y kt H k y k ) (1 + ρ vt y k sk u T y k s kt y k, β = ρ 1 v T y k. H ky k y kt H k y k This class of updates is known as the Broyden family. Note that if ρ = 0, we get the DFP update that we have seen before. Lemma 4 If ρ 0 then H k is positive definite for each k 0. Proof: It suffices if H k H ky k y kt H k y kt H k y k is positive definite, since the other two terms forming H k+1 are positive semidefinite. This however is an (almost) immediate consequence of the inequality of Cauchy-Schwarz. If ρ = 0 we get the DFP update. The choice ρ = 1 was proposed by Broyden, Fletcher, Goldfarb and Shanno (1970). It is the most popular BFGS update.. Optimization Group 31
32 Stopping criteria The stopping criterion is a relatively simple but essential part of all algorithms. If both primal and dual feasible solutions are generated then we use the duality gap primal obj. value dual obj. value as a criterion. We then stop the algorithm if the duality gap is smaller than some prescribed accuracy parameter ǫ. In unconstrained optimization one often uses a primal algorithm and then there is no such obvious measure for the distance to the optimum. We then stop if there is no sufficient improvement in the objective value, or if the iterates stay too close to each other or if the length of the gradient or the length of the Newton step in an appropriate norm is small. All these criteria can be scaled (relative to) some characteristic number describing the dimensions of the problem. For example, the relative improvement at two subsequent iterates x k, x k+1 in the objective value is usually measured by f(x k ) f(x k+1 ) 1 + f(x k, ) and we may stop if it smaller than a prescribed accuracy parameter ǫ. Optimization Group 32
Chapter 4. Unconstrained optimization
Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file
More informationQuasi-Newton methods for minimization
Quasi-Newton methods for minimization Lectures for PHD course on Numerical optimization Enrico Bertolazzi DIMS Universitá di Trento November 21 December 14, 2011 Quasi-Newton methods for minimization 1
More informationNonlinear Optimization: What s important?
Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global
More information5 Quasi-Newton Methods
Unconstrained Convex Optimization 26 5 Quasi-Newton Methods If the Hessian is unavailable... Notation: H = Hessian matrix. B is the approximation of H. C is the approximation of H 1. Problem: Solve min
More informationUnconstrained optimization
Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout
More informationMATH 4211/6211 Optimization Quasi-Newton Method
MATH 4211/6211 Optimization Quasi-Newton Method Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Quasi-Newton Method Motivation:
More informationOptimization and Root Finding. Kurt Hornik
Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428
More informationMethods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent
Nonlinear Optimization Steepest Descent and Niclas Börlin Department of Computing Science Umeå University niclas.borlin@cs.umu.se A disadvantage with the Newton method is that the Hessian has to be derived
More informationNonlinear Programming
Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week
More informationMethods for Unconstrained Optimization Numerical Optimization Lectures 1-2
Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Coralia Cartis, University of Oxford INFOMM CDT: Modelling, Analysis and Computation of Continuous Real-World Problems Methods
More information8 Numerical methods for unconstrained problems
8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields
More information1 Numerical optimization
Contents 1 Numerical optimization 5 1.1 Optimization of single-variable functions............ 5 1.1.1 Golden Section Search................... 6 1.1. Fibonacci Search...................... 8 1. Algorithms
More informationQuasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno February 6, / 25 (BFG. Limited memory BFGS (L-BFGS)
Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno (BFGS) Limited memory BFGS (L-BFGS) February 6, 2014 Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb
More informationOptimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23
Optimization: Nonlinear Optimization without Constraints Nonlinear Optimization without Constraints 1 / 23 Nonlinear optimization without constraints Unconstrained minimization min x f(x) where f(x) is
More information1 Numerical optimization
Contents Numerical optimization 5. Optimization of single-variable functions.............................. 5.. Golden Section Search..................................... 6.. Fibonacci Search........................................
More informationNewton s Method. Ryan Tibshirani Convex Optimization /36-725
Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x
More informationChapter 10 Conjugate Direction Methods
Chapter 10 Conjugate Direction Methods An Introduction to Optimization Spring, 2012 1 Wei-Ta Chu 2012/4/13 Introduction Conjugate direction methods can be viewed as being intermediate between the method
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 3 Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 2 3.1. Gradient method Classical gradient method: to minimize a differentiable convex
More informationLecture 7 Unconstrained nonlinear programming
Lecture 7 Unconstrained nonlinear programming Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University,
More informationConvex Optimization. Problem set 2. Due Monday April 26th
Convex Optimization Problem set 2 Due Monday April 26th 1 Gradient Decent without Line-search In this problem we will consider gradient descent with predetermined step sizes. That is, instead of determining
More informationStatistics 580 Optimization Methods
Statistics 580 Optimization Methods Introduction Let fx be a given real-valued function on R p. The general optimization problem is to find an x ɛ R p at which fx attain a maximum or a minimum. It is of
More informationQuasi-Newton Methods
Newton s Method Pros and Cons Quasi-Newton Methods MA 348 Kurt Bryan Newton s method has some very nice properties: It s extremely fast, at least once it gets near the minimum, and with the simple modifications
More informationFALL 2018 MATH 4211/6211 Optimization Homework 4
FALL 2018 MATH 4211/6211 Optimization Homework 4 This homework assignment is open to textbook, reference books, slides, and online resources, excluding any direct solution to the problem (such as solution
More informationConvex Optimization CMU-10725
Convex Optimization CMU-10725 Quasi Newton Methods Barnabás Póczos & Ryan Tibshirani Quasi Newton Methods 2 Outline Modified Newton Method Rank one correction of the inverse Rank two correction of the
More informationAlgorithms for Constrained Optimization
1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic
More informationGradient-Based Optimization
Multidisciplinary Design Optimization 48 Chapter 3 Gradient-Based Optimization 3. Introduction In Chapter we described methods to minimize (or at least decrease) a function of one variable. While problems
More information5 Handling Constraints
5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest
More informationAM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality
More informationStatic unconstrained optimization
Static unconstrained optimization 2 In unconstrained optimization an objective function is minimized without any additional restriction on the decision variables, i.e. min f(x) x X ad (2.) with X ad R
More informationLecture 14: October 17
1-725/36-725: Convex Optimization Fall 218 Lecture 14: October 17 Lecturer: Lecturer: Ryan Tibshirani Scribes: Pengsheng Guo, Xian Zhou Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:
More informationAM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α
More informationOptimization Methods
Optimization Methods Decision making Examples: determining which ingredients and in what quantities to add to a mixture being made so that it will meet specifications on its composition allocating available
More informationOptimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30
Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained
More informationOptimization II: Unconstrained Multivariable
Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Optimization II: Unconstrained Multivariable 1
More informationSECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS
SECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS HONOUR SCHOOL OF MATHEMATICS, OXFORD UNIVERSITY HILARY TERM 2005, DR RAPHAEL HAUSER 1. The Quasi-Newton Idea. In this lecture we will discuss
More informationLecture 10: September 26
0-725: Optimization Fall 202 Lecture 0: September 26 Lecturer: Barnabas Poczos/Ryan Tibshirani Scribes: Yipei Wang, Zhiguang Huo Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These
More informationThe Steepest Descent Algorithm for Unconstrained Optimization
The Steepest Descent Algorithm for Unconstrained Optimization Robert M. Freund February, 2014 c 2014 Massachusetts Institute of Technology. All rights reserved. 1 1 Steepest Descent Algorithm The problem
More information2. Quasi-Newton methods
L. Vandenberghe EE236C (Spring 2016) 2. Quasi-Newton methods variable metric methods quasi-newton methods BFGS update limited-memory quasi-newton methods 2-1 Newton method for unconstrained minimization
More informationNumerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09
Numerical Optimization 1 Working Horse in Computer Vision Variational Methods Shape Analysis Machine Learning Markov Random Fields Geometry Common denominator: optimization problems 2 Overview of Methods
More informationImproving the Convergence of Back-Propogation Learning with Second Order Methods
the of Back-Propogation Learning with Second Order Methods Sue Becker and Yann le Cun, Sept 1988 Kasey Bray, October 2017 Table of Contents 1 with Back-Propagation 2 the of BP 3 A Computationally Feasible
More informationOn nonlinear optimization since M.J.D. Powell
On nonlinear optimization since 1959 1 M.J.D. Powell Abstract: This view of the development of algorithms for nonlinear optimization is based on the research that has been of particular interest to the
More informationOptimization Tutorial 1. Basic Gradient Descent
E0 270 Machine Learning Jan 16, 2015 Optimization Tutorial 1 Basic Gradient Descent Lecture by Harikrishna Narasimhan Note: This tutorial shall assume background in elementary calculus and linear algebra.
More information4TE3/6TE3. Algorithms for. Continuous Optimization
4TE3/6TE3 Algorithms for Continuous Optimization (Algorithms for Constrained Nonlinear Optimization Problems) Tamás TERLAKY Computing and Software McMaster University Hamilton, November 2005 terlaky@mcmaster.ca
More informationWritten Examination
Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes
More informationMA/OR/ST 706: Nonlinear Programming Midterm Exam Instructor: Dr. Kartik Sivaramakrishnan INSTRUCTIONS
MA/OR/ST 706: Nonlinear Programming Midterm Exam Instructor: Dr. Kartik Sivaramakrishnan INSTRUCTIONS 1. Please write your name and student number clearly on the front page of the exam. 2. The exam is
More informationNumerical Methods I Solving Nonlinear Equations
Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)
More informationMultivariate Newton Minimanization
Multivariate Newton Minimanization Optymalizacja syntezy biosurfaktantu Rhamnolipid Rhamnolipids are naturally occuring glycolipid produced commercially by the Pseudomonas aeruginosa species of bacteria.
More informationE5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization
E5295/5B5749 Convex optimization with engineering applications Lecture 8 Smooth convex unconstrained and equality-constrained minimization A. Forsgren, KTH 1 Lecture 8 Convex optimization 2006/2007 Unconstrained
More informationOptimization II: Unconstrained Multivariable
Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Optimization II: Unconstrained
More informationConvex Optimization. Newton s method. ENSAE: Optimisation 1/44
Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)
More informationHigher-Order Methods
Higher-Order Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. PCMI, July 2016 Stephen Wright (UW-Madison) Higher-Order Methods PCMI, July 2016 1 / 25 Smooth
More informationComparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming Problems
International Journal of Scientific and Research Publications, Volume 3, Issue 10, October 013 1 ISSN 50-3153 Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming
More informationThe Conjugate Gradient Method
The Conjugate Gradient Method Lecture 5, Continuous Optimisation Oxford University Computing Laboratory, HT 2006 Notes by Dr Raphael Hauser (hauser@comlab.ox.ac.uk) The notion of complexity (per iteration)
More informationSecond Order Optimization Algorithms I
Second Order Optimization Algorithms I Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 7, 8, 9 and 10 1 The
More informationIE 5531: Engineering Optimization I
IE 5531: Engineering Optimization I Lecture 15: Nonlinear optimization Prof. John Gunnar Carlsson November 1, 2010 Prof. John Gunnar Carlsson IE 5531: Engineering Optimization I November 1, 2010 1 / 24
More informationJanuary 29, Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes Stiefel 1 / 13
Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière Hestenes Stiefel January 29, 2014 Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes
More informationOptimization Methods for Machine Learning
Optimization Methods for Machine Learning Sathiya Keerthi Microsoft Talks given at UC Santa Cruz February 21-23, 2017 The slides for the talks will be made available at: http://www.keerthis.com/ Introduction
More informationLecture V. Numerical Optimization
Lecture V Numerical Optimization Gianluca Violante New York University Quantitative Macroeconomics G. Violante, Numerical Optimization p. 1 /19 Isomorphism I We describe minimization problems: to maximize
More informationGradient Descent. Dr. Xiaowei Huang
Gradient Descent Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Three machine learning algorithms: decision tree learning k-nn linear regression only optimization objectives are discussed,
More informationNumerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen
Numerisches Rechnen (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang Institut für Geometrie und Praktische Mathematik RWTH Aachen Wintersemester 2011/12 IGPM, RWTH Aachen Numerisches Rechnen
More informationGlobal Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method. 1 Introduction
ISSN 1749-3889 (print), 1749-3897 (online) International Journal of Nonlinear Science Vol.11(2011) No.2,pp.153-158 Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method Yigui Ou, Jun Zhang
More informationUnconstrained Multivariate Optimization
Unconstrained Multivariate Optimization Multivariate optimization means optimization of a scalar function of a several variables: and has the general form: y = () min ( ) where () is a nonlinear scalar-valued
More informationECS550NFB Introduction to Numerical Methods using Matlab Day 2
ECS550NFB Introduction to Numerical Methods using Matlab Day 2 Lukas Laffers lukas.laffers@umb.sk Department of Mathematics, University of Matej Bel June 9, 2015 Today Root-finding: find x that solves
More informationPart 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)
Part 3: Trust-region methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted
More informationNonlinearOptimization
1/35 NonlinearOptimization Pavel Kordík Department of Computer Systems Faculty of Information Technology Czech Technical University in Prague Jiří Kašpar, Pavel Tvrdík, 2011 Unconstrained nonlinear optimization,
More informationInterior-Point Methods for Linear Optimization
Interior-Point Methods for Linear Optimization Robert M. Freund and Jorge Vera March, 204 c 204 Robert M. Freund and Jorge Vera. All rights reserved. Linear Optimization with a Logarithmic Barrier Function
More informationOptimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng
Optimization 2 CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Optimization 2 1 / 38
More informationNewton s Method. Javier Peña Convex Optimization /36-725
Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and
More informationOutline. Scientific Computing: An Introductory Survey. Optimization. Optimization Problems. Examples: Optimization Problems
Outline Scientific Computing: An Introductory Survey Chapter 6 Optimization 1 Prof. Michael. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction
More informationConvex Optimization M2
Convex Optimization M2 Lecture 3 A. d Aspremont. Convex Optimization M2. 1/49 Duality A. d Aspremont. Convex Optimization M2. 2/49 DMs DM par email: dm.daspremont@gmail.com A. d Aspremont. Convex Optimization
More informationMVE165/MMG631 Linear and integer optimization with applications Lecture 13 Overview of nonlinear programming. Ann-Brith Strömberg
MVE165/MMG631 Overview of nonlinear programming Ann-Brith Strömberg 2015 05 21 Areas of applications, examples (Ch. 9.1) Structural optimization Design of aircraft, ships, bridges, etc Decide on the material
More informationAlgorithms for constrained local optimization
Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained
More information(One Dimension) Problem: for a function f(x), find x 0 such that f(x 0 ) = 0. f(x)
Solving Nonlinear Equations & Optimization One Dimension Problem: or a unction, ind 0 such that 0 = 0. 0 One Root: The Bisection Method This one s guaranteed to converge at least to a singularity, i not
More informationScientific Computing: Optimization
Scientific Computing: Optimization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 March 8th, 2011 A. Donev (Courant Institute) Lecture
More informationReview of Classical Optimization
Part II Review of Classical Optimization Multidisciplinary Design Optimization of Aircrafts 51 2 Deterministic Methods 2.1 One-Dimensional Unconstrained Minimization 2.1.1 Motivation Most practical optimization
More informationLecture 15 Newton Method and Self-Concordance. October 23, 2008
Newton Method and Self-Concordance October 23, 2008 Outline Lecture 15 Self-concordance Notion Self-concordant Functions Operations Preserving Self-concordance Properties of Self-concordant Functions Implications
More informationCSCI 1951-G Optimization Methods in Finance Part 09: Interior Point Methods
CSCI 1951-G Optimization Methods in Finance Part 09: Interior Point Methods March 23, 2018 1 / 35 This material is covered in S. Boyd, L. Vandenberge s book Convex Optimization https://web.stanford.edu/~boyd/cvxbook/.
More informationNumerical solutions of nonlinear systems of equations
Numerical solutions of nonlinear systems of equations Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan E-mail: min@math.ntnu.edu.tw August 28, 2011 Outline 1 Fixed points
More informationOptimization for Machine Learning
Optimization for Machine Learning (Problems; Algorithms - A) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html
More informationarxiv: v3 [math.na] 23 Mar 2016
Randomized Quasi-Newton Updates are Linearly Convergent Matrix Inversion Algorithms Robert M. Gower and Peter Richtárik arxiv:602.0768v3 [math.na] 23 Mar 206 School of Mathematics University of Edinburgh
More informationNonlinear Optimization for Optimal Control
Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]
More information6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection
6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE Three Alternatives/Remedies for Gradient Projection Two-Metric Projection Methods Manifold Suboptimization Methods
More informationDO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO
QUESTION BOOKLET EECS 227A Fall 2009 Midterm Tuesday, Ocotober 20, 11:10-12:30pm DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO You have 80 minutes to complete the midterm. The midterm consists
More informationPrimal-Dual Interior-Point Methods for Linear Programming based on Newton s Method
Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Robert M. Freund March, 2004 2004 Massachusetts Institute of Technology. The Problem The logarithmic barrier approach
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationSuppose that the approximate solutions of Eq. (1) satisfy the condition (3). Then (1) if η = 0 in the algorithm Trust Region, then lim inf.
Maria Cameron 1. Trust Region Methods At every iteration the trust region methods generate a model m k (p), choose a trust region, and solve the constraint optimization problem of finding the minimum of
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted
More informationOptimization Methods
Optimization Methods Categorization of Optimization Problems Continuous Optimization Discrete Optimization Combinatorial Optimization Variational Optimization Common Optimization Concepts in Computer Vision
More informationChapter 8 Gradient Methods
Chapter 8 Gradient Methods An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Introduction Recall that a level set of a function is the set of points satisfying for some constant. Thus, a point
More informationLecture 18: November Review on Primal-dual interior-poit methods
10-725/36-725: Convex Optimization Fall 2016 Lecturer: Lecturer: Javier Pena Lecture 18: November 2 Scribes: Scribes: Yizhu Lin, Pan Liu Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:
More informationLecture 3: Linesearch methods (continued). Steepest descent methods
Lecture 3: Linesearch methods (continued). Steepest descent methods Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 3: Linesearch methods (continued).
More informationECE580 Fall 2015 Solution to Midterm Exam 1 October 23, Please leave fractions as fractions, but simplify them, etc.
ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, 2015 1 Name: Solution Score: /100 This exam is closed-book. You must show ALL of your work for full credit. Please read the questions carefully.
More informationOptimization Methods. Lecture 19: Line Searches and Newton s Method
15.93 Optimization Methods Lecture 19: Line Searches and Newton s Method 1 Last Lecture Necessary Conditions for Optimality (identifies candidates) x local min f(x ) =, f(x ) PSD Slide 1 Sufficient Conditions
More informationDetermination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study
International Journal of Mathematics And Its Applications Vol.2 No.4 (2014), pp.47-56. ISSN: 2347-1557(online) Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms:
More informationQuasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization
Quasi-Newton Methods Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization 10-725 Last time: primal-dual interior-point methods Given the problem min x f(x) subject to h(x)
More informationChapter 2. Optimization. Gradients, convexity, and ALS
Chapter 2 Optimization Gradients, convexity, and ALS Contents Background Gradient descent Stochastic gradient descent Newton s method Alternating least squares KKT conditions 2 Motivation We can solve
More informationOptimality Conditions for Constrained Optimization
72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)
More information