Nonlinear Programming

Size: px
Start display at page:

Download "Nonlinear Programming"

Transcription

1 Nonlinear Programming Kees Roos URL: roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D Optimization Group 1

2 Outline for week 7: Algorithms for unconstrained minimization A generic algorithm Rate of convergence Line search methods Dichotomous and golden section search Bisection Newton s method Search directions Gradient method Newton s method Methods of conjugate directions Powell s method Fletcher-Reeves method Quasi-Newton s method DFP update BFGS update Stopping criteria Optimization Group 2

3 Input: ǫ > 0 is the accuracy parameter; x 0 is a given (relative interior) feasible point; Generic algorithm for min x C f(x) Step 0: x := x 0, k = 0; Step 1: Find search direction s k s.t. δf(x k, s k ) < 0 (This should be a descending feasible direction in the constrained case.) Step 1a: If no such direction exists STOP, optimum found. Step 2: Line search : find λ k = argmin λ f(x k + λs k ); Step 3: x k+1 = x k + λ k s k, k = k + 1; Step 4: If stopping criteria satisfied STOP, else GOTO Step 1. Optimization Group 3

4 Algorithms: rate of convergence Definition: Let α 1, α 2,..., α k,... α be a convergent sequence. The rate of convergence is: p = sup { p : limsup k The larger p is, the faster the convergence. Let The rate of convergence is: β = limsup k α k+1 α α k α p < α k+1 α α k α p. }. linear if p = 1 and 0 < β < 1; super-linear if p = 1 and β = 0; quadratic if p = 2; sub-linear if β = 1. Optimization Group 4

5 Examples: order of convergence Example 1: The sequence α k = a k, where 0 < a < 1 converges linearly to zero while β = a. Example 2: The sequence α k = a (2k), where 0 < a < 1 converges quadratically to zero. Example 3: The sequence α k = 1 k converges sub-linearly to zero. Example 4: The sequence α k = ( 1 k )k converges super-linearly to zero. Optimization Group 5

6 Linesearch methods We assume throughout that f is a convex function. We are given a (feasible) search direction s at a feasible point x and we want to find So we are minimizing for λ 0. λ = argmin λ 0 f(x + λs). φ(λ) := f(x + λs) This is a one-dimensional problem. We deal with four different line search methods, that require different levels of information about φ(λ): The Dichotomous search and Golden section methods, that use only function evaluations of φ; Bisection, that evaluates φ (λ) (φ has to be continuously differentiable); Newton s method, that evaluates both φ (λ) and φ (λ). Optimization Group 6

7 Linesearch: Dichotomous search We assume that φ is convex and has a minimizer on the interval [a, b]. Our aim is to reduce the size of this interval of uncertainty by evaluating φ at points in [a, b]. Lemma 1 (Exercise 4.7) Let a < ā < b < b. If φ(ā) < φ( b) then the minimum of φ occurs in the interval [a, b]; if φ(ā) φ( b) the minimum of φ occurs in the interval [ā, b]. The lemma suggest a simple algorithm to reduce the interval of uncertainty. Optimization Group 7

8 Input: ǫ > 0 is the accuracy parameter; Linesearch: Dichotomous search a 0, b 0 are given such that [a 0, b 0 ] contains the minimizer of φ(λ), k = 0. Step 1: If a k b k < ǫ STOP. Step 2: Choose ā k (a k, b k ) and b k (a k, b k ), such that ā k < b k ; Step 3a: If φ(ā k ) < φ( b k ) set a k+1 = a k, b k+1 = b k ; Step 3b: If φ(ā k ) φ( b k ) set a k+1 = ā k, b k+1 = b k ; Step 4: Set k = k + 1. GOTO Step 1. We have not specified yet how we should choose the values ā k and b k in iteration k (Step 2 of the algorithm). There are many ways to do this. One is to choose ā k = 1 2 (a k + b k ) δ and b k = 1 2 (a k + b k ) + δ where δ > 0 is a (very) small fixed constant. Then the interval of uncertainty is reduced by a factor ( δ)t/2 after t function evaluations (Exercise 4.8). Optimization Group 8

9 Linesearch: Golden section search This is a variant of the Dichotomous search method where the constant δ is not constant but depends on k. In the k-th iteration we take δ = δ k, where δ k = ( α 1 2 ) (b k a k ), α = 1 2 ( 5 1 ) Here α is the Golden ratio, i.e., the root of α 2 + α 1 = 0 with α [0,1]. We now have ā k = 1 2 (a k + b k ) δ k = 1 2 (a k + b k ) ( α 1 2 (b k a k ) = b k α(b k a k ) b k = 1 2 (a k + b k ) + δ k = 1 2 (a k + b k ) + ( α 1 ) 2 (b k a k ) = a k + α(b k a k ). If φ(ā k ) < φ( b k ), then we set b k+1 = b k and a k+1 = a k. In that case b k+1 = a k+1 + α ( b k+1 a k+1 ) = ak + α ( b k a k ) = ak + α 2 (b k a k ) = a k + (1 α)(b k a k ) = b k α(b k a k ) = ā k. So in the next iteration we only need to compute φ(ā k+1 ). Similarly, if φ(ā k ) φ( b k ), then we set a k+1 = ā k and b k+1 = b k, and it follows in a similar way that ā k+1 = b k. So in the next iteration we only need to compute φ( b k+1 ). In both cases one need to evaluate φ only once. See the course notes for graphical illustrations. When using Golden section search each iteration reduces the interval of uncertainty by a factor α (Exercise 4.9). ) Optimization Group 9

10 Linesearch: Golden section search Suppose a k = 0 and b k = 1. We choose fixed α ( 1 2,1) and define ā k+1 = a k + (1 α)(b k a k ) = 1 α, b k+1 = b k (1 α)(b k a k ) = α. Suppose φ(1 α) < φ(α). Then we set a k+1 = 0 and b k+1 = α and ā k+2 = a k+1 + (1 α)(b k+1 a k+1 ) = (1 α)α b k+2 = b k+1 (1 α)(b k+1 a k+1 ) = α (1 α)α. We want one of these two points to be ā k+1 = 1 α, because we already know φ(1 α). This gives either or, equivalently, (1 α)α = 1 α or α α(1 α) = 1 α, α = 1 or α 2 + α 1 = 0. Since α ( 1 2,1), the only possible value for α is which is the Golden ratio! α = 1 2 ( 5 1 ) 0.618, Optimization Group 10

11 Linesearch: Bisection (or Bolzano s method) We assume that φ(λ) is differentiable (and convex). We wish to find a λ so that φ ( λ) = 0. Input: ǫ > 0 is the accuracy parameter; a 0, b 0 are given such that φ (a 0 ) < 0 & φ (b 0 ) > 0, k = 0; Step 1: If b k a k < ǫ STOP. Step 2: Let λ = 1 2 (a k + b k ); Step 3a: If φ (λ) < 0 then set a k+1 = λ, b k+1 = b k ; Step 3b: If φ (λ) > 0 then set a k+1 = a k, b k+1 = λ; Step 4: Set k = k + 1. GOTO Step 1. The algorithm needs log 2 b 0 a 0 ǫ evaluations of φ (Exercise 4.11). Optimization Group 11

12 Linesearch using the Newton-Raphson method The quadratic approximation of φ at λ k is q(λ) = φ(λ k ) + φ (λ k )(λ λ k ) φ (λ k )(λ λ k ) 2. The minimum of q is attained if q (λ) = 0, which gives Input: ǫ > 0 is the accuracy parameter; λ 0 is the given initial point; k = 0; λ k+1 = λ k φ (λ k ) φ (λ k ). Step 1: Let λ k+1 = λ k φ (λ k ) φ (λ k ) ; Step 2: If λ k+1 λ k < ǫ STOP. Step 3: k := k + 1, GOTO Step 1. Optimization Group 12

13 The Newton-Raphson method: Example 4.3 Let φ(λ) = λ log(1+λ). The domain of φ is ( 1, ). The first and second derivatives of φ are φ (λ) = λ 1 + λ, φ (λ) = 1 (1 + λ) 2. This makes clear that φ is strictly convex on its domain, and minimal at λ = 0. The iterates satisfy the recursive relation λ k+1 = λ k φ (λ k ) φ (λ k ) = λ k λ k (1 + λ k ) = λ 2 k. This implies quadratic convergence if λ 0 < 1 (see Exercise 4.12). On the other hand, Newton s method fails if λ 0 1. For example, if λ 0 = 1 then λ 1 = 1, which is not in the domain of φ! In general the method converges quadratically if the following conditions are met: 1. the starting point is sufficiently close to the minimizer; 2. in addition to being convex, the function φ has a property called self-concordance, which is introduced later. Optimization Group 13

14 Search Directions: The Gradient method Search direction: s = f(x k ) Steepest descent direction! δf(x, f(x)) = f(x) T f(x) = min s = f(x) { f(x) T s}. The (negative) gradient is orthogonal to the level curves (Exercise 4.14). The gradient method is not a finite algorithm, not even for linear or quadratic functions. Slow convergence ( zigzagging ) (Figure 4.4). The order of convergence is only linear. Optimization Group 14

15 Let f be continuously differentiable. Convergence of the gradient method Start from the initial point x 0 using exact line search; the gradient method produces a sequence { x k} such that f(x k ) > f(x k+1 ) for k = 0,1,2,. Assume that the level set D = {x : f(x) f(x 0 )} is compact, then any accumulation point x of the sequence { x k} is a stationary point (i.e. f( x) = 0) of f. If the function f is a convex function, then x is a global minimizer of f. If the function f is not convex, then x is a local minimizer of f. Optimization Group 15

16 Newton s method Newton s method is based on minimizing the second order approximation of f at x k. q(x) := f(x k ) + f(x k ) T (x x k ) (x xk ) T 2 f(x k )(x x k ). We assume that q(x) is strictly convex. So the Hessian 2 f(x k ) is positive definite. Hence the minimum is attained when q(x) = f(x k ) + 2 f(x k )(x x k ) = 0. We can solve x from 2 f(x k )(x x k ) = f(x k ) which gives the next iterate: x k+1 = x k ( 2 f(x k )) 1 f(x k ), So the Newton direction is s k = ( 2 f(x k )) 1 f(x k ). Exact when f is quadratic. Local quadratic convergence with full Newton steps (α = 1, so without any line search!). Good starting point is essential. Optimization Group 16

17 Trust region method If the function f(x) is not strictly convex, or if the Hessian is ill-conditioned, then the Hessian is not (or hardly) invertible. Remedy: trust region method. 2 f(x) is replaced by ( 2 f(x) + αi); s k = ( 2 f(x k ) + αi) 1 f(x k ). α is dynamically increased and decreased in order to avoid exact line search. If α = 0 then we have the Newton step, if α then we approach a (small) multiple of the negative gradient. Optimization Group 17

18 Newton s method for solving nonlinear equations Find solution of F(x) = 0, where F : R m R n Linearize at x k : F(x) F(x k ) + JF(x k )(x x k ) Jacobian of F: JF(x) ij = F i(x) x j. Solve x k+1 from JF(x k )(x k+1 x k ) = F(x k ). Minimizing f(x) is equivalent to solving f(x) = 0. 2 f(x k )(x k+1 x k ) = f(x k ). The Jacobian of the gradient is exactly the Hessian of the function f(x) hence it is positive definite and we have as we have seen above. x k+1 = x k ( 2 f(x k )) 1 f(x k ) Conclusion: Newton s optimization method Newton s method for nonlinear equations applied to the system f(x) = 0. Optimization Group 18

19 Methods using conjugate directions (1) Let A be an n n symmetric positive definite matrix and b R n. We consider min { q(x) = 1 2 xt Ax b T x : x R n}. The minimizer is uniquely determined by q(x) = Ax b = 0. But to find the minimizer we need to invert the matrix A. If n is large this is computationally expensive, and we want to avoid this. This can be done by using so-called conjugate search directions. If the subsequent search directions are s 0,..., s k then the iterates have the form x k+1 = x k + λ k s k, k = 1, 2,.... If we use exact line search, then we automatically have q(x k+1 ) T s k = 0, k = 1, 2,.... By requiring a little more, namely that the search vectors s i are linearly independent and q(x k+1 ) T s i = 0, 0 i k = 1, 2,... we can guarantee termination of the algorithm in a finite number of steps. Because then q(x n ) T s i = 0, i < n, whence, since the vectors s i are linearly independent, q(x n ) = 0. So no more than n steps are required. Optimization Group 19

20 Methods using conjugate directions (2) We denote q(x k ) as g k. Note that when using exact line search, we have (automatically) g j+1t s j = 0, j = 1, 2,.... Lemma 2 Let k {1,..., n}. The following two statement are equivalent: (i) g j+1t s i = 0, 0 i < j k; (ii) s it As j = 0, 0 i < j k. Proof: Since q(x) = Ax b, we have q(x j+1 ) = Ax j+1 b = A ( x j + λ j s j) b = q(x j ) + λ j As j = g j + λ j As j, j = 0, 1,... Therefore, for each i 0, g j+1t s i = q(x j+1 ) T s i = g jt s i + λ j s it As j, j = 0, 1,.... The proof can now be easily completed by induction on k, since λ j > 0 for each j. If (ii) holds then the vectors s 0,..., s k R n are called conjugate (or A conjugate). Note that if A = I then conjugate means orthogonal. Then s 0,..., s k are linearly independent. This also holds for A-conjugate vectors, since A is positive definite (Ex. 4.20). As we established before, if one uses A-conjugate directions to minimize the quadratic form q, then the minimizer of q is found in at most n iterations. Optimization Group 20

21 Easy method to generate conjugate directions Let s 0 = q(x 0 ) = g 0. Then we can get subsequent conjugate directions by taking s k = g k + α k s k 1, k = 1, 2,... for suitable values of α k. In order to make s k and s k 1 A-conjugate, we must have s kt As k 1 = 0 for k 1. This determines the coefficients α k uniquely: α k = gkt As k 1 k 1. s k 1T Ask 1, We proceed with induction on k. So we assume that s 0,..., s k 1 are conjugate. Using g kt s k 1 = 0 we find g kt s k = g kt ( g k + α k s k 1) = g k 2 < 0, proving that s k is a descent direction, provided g k 0. Our choice of α k implies s kt As k 1 = 0. So it remains to show that s kt As i = 0 for i < k 1. The induction hypothesis implies s kt As i = ( g k + α k s k 1) T As i = g kt As i. Since g i = q(x i ) = Ax i b, and x i+1 = x i + λ i s i, we have Hence, due to Lemma 2(ii), λ i As i = g i+1 g i = α i+1 s i s i+1 ( α i s i 1 s i). λ i g kt As i = g kt ( αi+1 s i s i+1 α i s i 1 + s i) = 0. This proves that s 0,..., s k are conjugate, provided g k 0 (otherwise x k is optimal!). Optimization Group 21

22 The case of nonquadratic functions In the case where f is (convex) quadratic then finite termination is guaranteed if ( ) ( ) α k = gkt As k 1 s k 1T As = gkt gk gk 1 k 1 s k 1T (g k g k 1 ) = gkt gk gk 1 g k 2 g k 1 2 = g k 1 2, k 1. Here we used that g k g k 1, as follows from g k 1 = s k 1 +α k s k 2 and g kt s k 2 = 0, by Lemma 2 (i) and g kt s k 1 = 0 by the choice of λ k. The algorithm is Step 0. Let s 0 = f(x 0 ) and x 1 := argmin f(x 0 + λs 0 ). Step k. Set s k = f(x k ) + α k s k 1 and x k+1 := argmin f(x k + λs k ). If f is not quadratic there is no guarantee that the method stops after a finite number of steps. Several choices for α k have been proposed (which are equivalent in the quadratic case): ( Hestenes-Stiefel (1952): α k = gkt g k g k 1) s k 1T (g k g k 1 ). g k 2 Fletcher-Reeves (1964): α k = g k 1 2. Polak-Ribière (1969): α k = gkt ( g k g k 1) g k 1 2. Optimization Group 22

23 Solving a linear system with the conjugate gradient method Assume we want to solve Ax = b with A positive definite. The solution is precisely the minimizer of q(x) = 1 2 xt Ax b T x, and hence can be solve by the conjugate gradient method. If A is not positive definite, but nonsingular, then A T A is positive definite. Hence we can solve Ax = b by minimizing q(x) = Ax b 2 = x T A T Ax 2b T Ax + b T b. Optimization Group 23

24 Powell s method We now deal with a conjugate direction method using only function values (no gradients!). Input A starting point x 0, a set of linearly independent vectors t 1,..., t n. Initialization Set t (1,i) = t i, i = 1,..., n. For k = 1,..., n do (Cycle k) Let z (k,1) = x k 1 and z (k,i+1) := argmin q ( z (k,i) + λt (k,i)), i = 1,..., n. Let x k := argmin q(z (k,n+1) + λs k ), where s k := z (k,n+1) x k 1. Let t (k+1,i) = t (k,i+1), i = 1,, n 1 and t (k+1,n) := s k. The algorithm consists of n cycles and terminates at the minimizer of q(x). Each cycle consists of n + 1 line searches and yields a search direction s k. The k-th direction s k is constructed at the end of cycle k. The search directions s 1,..., s n are conjugate (for a proof: see the course notes). Note that only function values are evaluated (no derivatives are used, unless the line searches do so). The number of line searches is n(n + 1). Therefore, Powell s method is attractive for minimizing black box functions where gradient and Hessian information is not available (or too expensive to compute). Optimization Group 24

25 Illustration of Powell s method 2 x t x 1 x 2 0 x 2 optimal t x 1 Iterates of Powell s algorithm for f(x) = 5x x 1x 2 + x , starting at x0 = [1;2]. Optimization Group 25

26 Quasi-Newton methods Recall that the Newton direction at iteration k is given by: s k = [ 2 f(x k ) ] 1 f(x k ) = [ 2 f(x k ) ] 1 g k. Quasi-Newton methods use a positive definite approximation H k to [ 2 f(x k ) ] 1. The approximation H k is updated at each iteration, say H k+1 = H k + D k, where D k denotes the update. The algorithm has the following generic form. Step 0. Let x 0 be given and set H 0 = I. Step k. s k = H k g k and x k+1 = argmin λ f(x k + λs k ) = x k + λ k s k ; H k+1 = H k + D k and k = k + 1. Defining we require for each k that y k := g k+1 g k, σ k = x k+1 x k = λ k s k. I H k+1 is symmetric positive definite; II σ k = H k+1 y k (quasi-newton property); III σ i = H k+1 y i (i = 0,..., k 1) (hereditary property), Optimization Group 26

27 The quasi-newton property and hereditary property Let A be an n n symmetric PD matrix, and let f be the strictly convex quadratic function Then g k = q(x k ) = Ax k b, and hence q(x) = 1 2 xt Ax b T x. y k = g k+1 g k = q(x k+1 ) q(x k ) = A ( x k+1 x k) = Aσ k, whence σ k = A 1 y k. Recall that each H k should be a good approximation of the inverse of 2 q(x k ), which is A 1. Therefore we require that σ k = H k+1 y k, which is the quasi-newton property, and even more, that our approximation H k satisfies σ i = H k+1 y i, i = 0,..., k. which is the hereditary property. Note that the hereditary property implies σ i = H n y i, i = 0,..., n 1. If the σ i (i = 0,..., n 1) are linearly independent, this implies H n = A 1. But then the n + 1-th iteration is simply the Newton step at x n. Since q is quadratic, this yields the minimizer of q, and hence we find the minimum of q no more than n + 1 iterations. Optimization Group 27

28 A generic update D k (1) First consider the case where D k is a (possibly indefinite) matrix of rank 2, whence D k = αuu T + βvv T for suitable vectors u and v and scalars α, β. Then the quasi-newton property implies H k+1 y k = H k y k + αuu T y k + βvv T y k = σ k. Davidon, Fletcher and Powell (1963) recognized that this condition is satisfied if u = σ k = λ k s k, α = 1, v = H k y k 1, β = σ kt y k y kt H k y k, which yields the so-called DFP update: D k = λ k s k s kt s kt y k H kyk ykt Hk y kt H k y k. In the following we consider a slight more general update, namely D k = αuu T + βvv T + µ ( uv T + vu T ) = [ ] u v α µ µ β Exercise A Show that D k has rank at most 2. ut v T. Optimization Group 28

29 A generic update D k (2) D k = αuu T + βvv T + µ ( uv T + vu T) = [ u v ] [ α µ µ β ] [ u T v T ], u = σ k, v = H k y k. Lemma 3 If the above update D k satisfies the quasi-newton property then the subsequent directions are conjugate. (So a quasi-newton method is a conjugate gradient method!) Proof: We show by induction on k that H k y i = σ i = λ i s i, s kt As i = g kt s i = 0, 0 i < k. (1) This trivially holds if k = 0 (the condition is void). Assuming the quasi-newton property and (1) for k 0, and using y i = Aσ i and σ i = λ i s i for all i, we write for i < k: y kt H k+1 y i = σ kt y i = σ kt Aσ i = λ k λ i s kt As i = 0. Also σ T k y i = σ T k Aσ i = 0. Hence we obtain for all i < k, D k y i = [ u v ] [ ] [ ] α µ u T y i µ β v T = [ σ k H k y ] [ k α µ y i µ β ] [ σ kt y i y kt H k y i whence H k+1 y i = H k y i +D k y i = H k y i = σ i. Together with the quasi-newton property this gives H k+1 y i = σ i for 0 i < k + 1. Because λ i 0, s k+1 = H k+1 g k+1 and H k+1 y i = σ i we observe next that λ i s k+1t As i = s k+1t Aσ i = s k+1t y i = g k+1t H k+1 y i = g k+1t σ i. Hence it suffices for the rest of the proof if g k+1t s i = 0 for 0 i < k + 1. This certainly holds if i = k, because we use exact line search. For i < k we use the induction hypothesis again, and g k+1 = A(x k + λ k s k ) b = g k + λ k As k, which gives This completes the proof. g k+1t s i = g kt s i + λ k s kt As = 0. ] = 0, Optimization Group 29

30 The Broyden family of updates (1) D k = αuu T + βvv T + µ ( uv T + vu T) = [ u v ] [ α µ µ β ] [ u T v T ], u = σ k, v = H k y k. We now determine conditions on the parameters α, β and µ that guarantee that each H k is positive definite. The quasi-newton property (σ k = H k+1 y k = H k y k + D k y k for k 0) amounts to u = v + αuu T y k + βvv T y k + µ ( uv T + vu T ) y k To satisfy this condition it suffices if = v ( 1 + βv T y k + µu T y k) + u ( αu T y k + µv T y k). αu T y k + µv T y k = 1 βv T y k + µu T y k = 1. This linear system has multiple solutions. Introducing ρ = µu T y k R the solution is α = 1 u T y k Since v T y k = y T H k y k > 0 and ( 1 + ρ vt y k u T y k ), β = ρ 1 v T y k, ρ R. u T y k = σ k T ( g k+1 g k) = λ k s k T ( g k+1 g k) = λ k s k T g k = λ k g k T H k g k > 0, the above expressions are well defined. Optimization Group 30

31 The Broyden family of updates (2) D k = αuu T +βvv T +µ ( uv T + vu T), u = σ k, v = H k y k, α = 1 u T y k Substituting the values of u, v, α and β we find D k = λ k s k s kt s kt y k H kyk ykt Hk y kt H k y k +ρwwt, w = y kt H k y k ) (1 + ρ vt y k sk u T y k s kt y k, β = ρ 1 v T y k. H ky k y kt H k y k This class of updates is known as the Broyden family. Note that if ρ = 0, we get the DFP update that we have seen before. Lemma 4 If ρ 0 then H k is positive definite for each k 0. Proof: It suffices if H k H ky k y kt H k y kt H k y k is positive definite, since the other two terms forming H k+1 are positive semidefinite. This however is an (almost) immediate consequence of the inequality of Cauchy-Schwarz. If ρ = 0 we get the DFP update. The choice ρ = 1 was proposed by Broyden, Fletcher, Goldfarb and Shanno (1970). It is the most popular BFGS update.. Optimization Group 31

32 Stopping criteria The stopping criterion is a relatively simple but essential part of all algorithms. If both primal and dual feasible solutions are generated then we use the duality gap primal obj. value dual obj. value as a criterion. We then stop the algorithm if the duality gap is smaller than some prescribed accuracy parameter ǫ. In unconstrained optimization one often uses a primal algorithm and then there is no such obvious measure for the distance to the optimum. We then stop if there is no sufficient improvement in the objective value, or if the iterates stay too close to each other or if the length of the gradient or the length of the Newton step in an appropriate norm is small. All these criteria can be scaled (relative to) some characteristic number describing the dimensions of the problem. For example, the relative improvement at two subsequent iterates x k, x k+1 in the objective value is usually measured by f(x k ) f(x k+1 ) 1 + f(x k, ) and we may stop if it smaller than a prescribed accuracy parameter ǫ. Optimization Group 32

Chapter 4. Unconstrained optimization

Chapter 4. Unconstrained optimization Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file

More information

Quasi-Newton methods for minimization

Quasi-Newton methods for minimization Quasi-Newton methods for minimization Lectures for PHD course on Numerical optimization Enrico Bertolazzi DIMS Universitá di Trento November 21 December 14, 2011 Quasi-Newton methods for minimization 1

More information

Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

More information

5 Quasi-Newton Methods

5 Quasi-Newton Methods Unconstrained Convex Optimization 26 5 Quasi-Newton Methods If the Hessian is unavailable... Notation: H = Hessian matrix. B is the approximation of H. C is the approximation of H 1. Problem: Solve min

More information

Unconstrained optimization

Unconstrained optimization Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout

More information

MATH 4211/6211 Optimization Quasi-Newton Method

MATH 4211/6211 Optimization Quasi-Newton Method MATH 4211/6211 Optimization Quasi-Newton Method Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Quasi-Newton Method Motivation:

More information

Optimization and Root Finding. Kurt Hornik

Optimization and Root Finding. Kurt Hornik Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428

More information

Methods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent

Methods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent Nonlinear Optimization Steepest Descent and Niclas Börlin Department of Computing Science Umeå University niclas.borlin@cs.umu.se A disadvantage with the Newton method is that the Hessian has to be derived

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week

More information

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Coralia Cartis, University of Oxford INFOMM CDT: Modelling, Analysis and Computation of Continuous Real-World Problems Methods

More information

8 Numerical methods for unconstrained problems

8 Numerical methods for unconstrained problems 8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields

More information

1 Numerical optimization

1 Numerical optimization Contents 1 Numerical optimization 5 1.1 Optimization of single-variable functions............ 5 1.1.1 Golden Section Search................... 6 1.1. Fibonacci Search...................... 8 1. Algorithms

More information

Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno February 6, / 25 (BFG. Limited memory BFGS (L-BFGS)

Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno February 6, / 25 (BFG. Limited memory BFGS (L-BFGS) Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno (BFGS) Limited memory BFGS (L-BFGS) February 6, 2014 Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb

More information

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23 Optimization: Nonlinear Optimization without Constraints Nonlinear Optimization without Constraints 1 / 23 Nonlinear optimization without constraints Unconstrained minimization min x f(x) where f(x) is

More information

1 Numerical optimization

1 Numerical optimization Contents Numerical optimization 5. Optimization of single-variable functions.............................. 5.. Golden Section Search..................................... 6.. Fibonacci Search........................................

More information

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Newton s Method. Ryan Tibshirani Convex Optimization /36-725 Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x

More information

Chapter 10 Conjugate Direction Methods

Chapter 10 Conjugate Direction Methods Chapter 10 Conjugate Direction Methods An Introduction to Optimization Spring, 2012 1 Wei-Ta Chu 2012/4/13 Introduction Conjugate direction methods can be viewed as being intermediate between the method

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 3 Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 2 3.1. Gradient method Classical gradient method: to minimize a differentiable convex

More information

Lecture 7 Unconstrained nonlinear programming

Lecture 7 Unconstrained nonlinear programming Lecture 7 Unconstrained nonlinear programming Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University,

More information

Convex Optimization. Problem set 2. Due Monday April 26th

Convex Optimization. Problem set 2. Due Monday April 26th Convex Optimization Problem set 2 Due Monday April 26th 1 Gradient Decent without Line-search In this problem we will consider gradient descent with predetermined step sizes. That is, instead of determining

More information

Statistics 580 Optimization Methods

Statistics 580 Optimization Methods Statistics 580 Optimization Methods Introduction Let fx be a given real-valued function on R p. The general optimization problem is to find an x ɛ R p at which fx attain a maximum or a minimum. It is of

More information

Quasi-Newton Methods

Quasi-Newton Methods Newton s Method Pros and Cons Quasi-Newton Methods MA 348 Kurt Bryan Newton s method has some very nice properties: It s extremely fast, at least once it gets near the minimum, and with the simple modifications

More information

FALL 2018 MATH 4211/6211 Optimization Homework 4

FALL 2018 MATH 4211/6211 Optimization Homework 4 FALL 2018 MATH 4211/6211 Optimization Homework 4 This homework assignment is open to textbook, reference books, slides, and online resources, excluding any direct solution to the problem (such as solution

More information

Convex Optimization CMU-10725

Convex Optimization CMU-10725 Convex Optimization CMU-10725 Quasi Newton Methods Barnabás Póczos & Ryan Tibshirani Quasi Newton Methods 2 Outline Modified Newton Method Rank one correction of the inverse Rank two correction of the

More information

Algorithms for Constrained Optimization

Algorithms for Constrained Optimization 1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic

More information

Gradient-Based Optimization

Gradient-Based Optimization Multidisciplinary Design Optimization 48 Chapter 3 Gradient-Based Optimization 3. Introduction In Chapter we described methods to minimize (or at least decrease) a function of one variable. While problems

More information

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality

More information

Static unconstrained optimization

Static unconstrained optimization Static unconstrained optimization 2 In unconstrained optimization an objective function is minimized without any additional restriction on the decision variables, i.e. min f(x) x X ad (2.) with X ad R

More information

Lecture 14: October 17

Lecture 14: October 17 1-725/36-725: Convex Optimization Fall 218 Lecture 14: October 17 Lecturer: Lecturer: Ryan Tibshirani Scribes: Pengsheng Guo, Xian Zhou Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α

More information

Optimization Methods

Optimization Methods Optimization Methods Decision making Examples: determining which ingredients and in what quantities to add to a mixture being made so that it will meet specifications on its composition allocating available

More information

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained

More information

Optimization II: Unconstrained Multivariable

Optimization II: Unconstrained Multivariable Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Optimization II: Unconstrained Multivariable 1

More information

SECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS

SECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS SECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS HONOUR SCHOOL OF MATHEMATICS, OXFORD UNIVERSITY HILARY TERM 2005, DR RAPHAEL HAUSER 1. The Quasi-Newton Idea. In this lecture we will discuss

More information

Lecture 10: September 26

Lecture 10: September 26 0-725: Optimization Fall 202 Lecture 0: September 26 Lecturer: Barnabas Poczos/Ryan Tibshirani Scribes: Yipei Wang, Zhiguang Huo Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These

More information

The Steepest Descent Algorithm for Unconstrained Optimization

The Steepest Descent Algorithm for Unconstrained Optimization The Steepest Descent Algorithm for Unconstrained Optimization Robert M. Freund February, 2014 c 2014 Massachusetts Institute of Technology. All rights reserved. 1 1 Steepest Descent Algorithm The problem

More information

2. Quasi-Newton methods

2. Quasi-Newton methods L. Vandenberghe EE236C (Spring 2016) 2. Quasi-Newton methods variable metric methods quasi-newton methods BFGS update limited-memory quasi-newton methods 2-1 Newton method for unconstrained minimization

More information

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09 Numerical Optimization 1 Working Horse in Computer Vision Variational Methods Shape Analysis Machine Learning Markov Random Fields Geometry Common denominator: optimization problems 2 Overview of Methods

More information

Improving the Convergence of Back-Propogation Learning with Second Order Methods

Improving the Convergence of Back-Propogation Learning with Second Order Methods the of Back-Propogation Learning with Second Order Methods Sue Becker and Yann le Cun, Sept 1988 Kasey Bray, October 2017 Table of Contents 1 with Back-Propagation 2 the of BP 3 A Computationally Feasible

More information

On nonlinear optimization since M.J.D. Powell

On nonlinear optimization since M.J.D. Powell On nonlinear optimization since 1959 1 M.J.D. Powell Abstract: This view of the development of algorithms for nonlinear optimization is based on the research that has been of particular interest to the

More information

Optimization Tutorial 1. Basic Gradient Descent

Optimization Tutorial 1. Basic Gradient Descent E0 270 Machine Learning Jan 16, 2015 Optimization Tutorial 1 Basic Gradient Descent Lecture by Harikrishna Narasimhan Note: This tutorial shall assume background in elementary calculus and linear algebra.

More information

4TE3/6TE3. Algorithms for. Continuous Optimization

4TE3/6TE3. Algorithms for. Continuous Optimization 4TE3/6TE3 Algorithms for Continuous Optimization (Algorithms for Constrained Nonlinear Optimization Problems) Tamás TERLAKY Computing and Software McMaster University Hamilton, November 2005 terlaky@mcmaster.ca

More information

Written Examination

Written Examination Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes

More information

MA/OR/ST 706: Nonlinear Programming Midterm Exam Instructor: Dr. Kartik Sivaramakrishnan INSTRUCTIONS

MA/OR/ST 706: Nonlinear Programming Midterm Exam Instructor: Dr. Kartik Sivaramakrishnan INSTRUCTIONS MA/OR/ST 706: Nonlinear Programming Midterm Exam Instructor: Dr. Kartik Sivaramakrishnan INSTRUCTIONS 1. Please write your name and student number clearly on the front page of the exam. 2. The exam is

More information

Numerical Methods I Solving Nonlinear Equations

Numerical Methods I Solving Nonlinear Equations Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)

More information

Multivariate Newton Minimanization

Multivariate Newton Minimanization Multivariate Newton Minimanization Optymalizacja syntezy biosurfaktantu Rhamnolipid Rhamnolipids are naturally occuring glycolipid produced commercially by the Pseudomonas aeruginosa species of bacteria.

More information

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization E5295/5B5749 Convex optimization with engineering applications Lecture 8 Smooth convex unconstrained and equality-constrained minimization A. Forsgren, KTH 1 Lecture 8 Convex optimization 2006/2007 Unconstrained

More information

Optimization II: Unconstrained Multivariable

Optimization II: Unconstrained Multivariable Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Optimization II: Unconstrained

More information

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44 Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)

More information

Higher-Order Methods

Higher-Order Methods Higher-Order Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. PCMI, July 2016 Stephen Wright (UW-Madison) Higher-Order Methods PCMI, July 2016 1 / 25 Smooth

More information

Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming Problems

Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming Problems International Journal of Scientific and Research Publications, Volume 3, Issue 10, October 013 1 ISSN 50-3153 Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Lecture 5, Continuous Optimisation Oxford University Computing Laboratory, HT 2006 Notes by Dr Raphael Hauser (hauser@comlab.ox.ac.uk) The notion of complexity (per iteration)

More information

Second Order Optimization Algorithms I

Second Order Optimization Algorithms I Second Order Optimization Algorithms I Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 7, 8, 9 and 10 1 The

More information

IE 5531: Engineering Optimization I

IE 5531: Engineering Optimization I IE 5531: Engineering Optimization I Lecture 15: Nonlinear optimization Prof. John Gunnar Carlsson November 1, 2010 Prof. John Gunnar Carlsson IE 5531: Engineering Optimization I November 1, 2010 1 / 24

More information

January 29, Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes Stiefel 1 / 13

January 29, Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes Stiefel 1 / 13 Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière Hestenes Stiefel January 29, 2014 Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes

More information

Optimization Methods for Machine Learning

Optimization Methods for Machine Learning Optimization Methods for Machine Learning Sathiya Keerthi Microsoft Talks given at UC Santa Cruz February 21-23, 2017 The slides for the talks will be made available at: http://www.keerthis.com/ Introduction

More information

Lecture V. Numerical Optimization

Lecture V. Numerical Optimization Lecture V Numerical Optimization Gianluca Violante New York University Quantitative Macroeconomics G. Violante, Numerical Optimization p. 1 /19 Isomorphism I We describe minimization problems: to maximize

More information

Gradient Descent. Dr. Xiaowei Huang

Gradient Descent. Dr. Xiaowei Huang Gradient Descent Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Three machine learning algorithms: decision tree learning k-nn linear regression only optimization objectives are discussed,

More information

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen Numerisches Rechnen (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang Institut für Geometrie und Praktische Mathematik RWTH Aachen Wintersemester 2011/12 IGPM, RWTH Aachen Numerisches Rechnen

More information

Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method. 1 Introduction

Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method. 1 Introduction ISSN 1749-3889 (print), 1749-3897 (online) International Journal of Nonlinear Science Vol.11(2011) No.2,pp.153-158 Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method Yigui Ou, Jun Zhang

More information

Unconstrained Multivariate Optimization

Unconstrained Multivariate Optimization Unconstrained Multivariate Optimization Multivariate optimization means optimization of a scalar function of a several variables: and has the general form: y = () min ( ) where () is a nonlinear scalar-valued

More information

ECS550NFB Introduction to Numerical Methods using Matlab Day 2

ECS550NFB Introduction to Numerical Methods using Matlab Day 2 ECS550NFB Introduction to Numerical Methods using Matlab Day 2 Lukas Laffers lukas.laffers@umb.sk Department of Mathematics, University of Matej Bel June 9, 2015 Today Root-finding: find x that solves

More information

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL) Part 3: Trust-region methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted

More information

NonlinearOptimization

NonlinearOptimization 1/35 NonlinearOptimization Pavel Kordík Department of Computer Systems Faculty of Information Technology Czech Technical University in Prague Jiří Kašpar, Pavel Tvrdík, 2011 Unconstrained nonlinear optimization,

More information

Interior-Point Methods for Linear Optimization

Interior-Point Methods for Linear Optimization Interior-Point Methods for Linear Optimization Robert M. Freund and Jorge Vera March, 204 c 204 Robert M. Freund and Jorge Vera. All rights reserved. Linear Optimization with a Logarithmic Barrier Function

More information

Optimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng

Optimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng Optimization 2 CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Optimization 2 1 / 38

More information

Newton s Method. Javier Peña Convex Optimization /36-725

Newton s Method. Javier Peña Convex Optimization /36-725 Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and

More information

Outline. Scientific Computing: An Introductory Survey. Optimization. Optimization Problems. Examples: Optimization Problems

Outline. Scientific Computing: An Introductory Survey. Optimization. Optimization Problems. Examples: Optimization Problems Outline Scientific Computing: An Introductory Survey Chapter 6 Optimization 1 Prof. Michael. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

More information

Convex Optimization M2

Convex Optimization M2 Convex Optimization M2 Lecture 3 A. d Aspremont. Convex Optimization M2. 1/49 Duality A. d Aspremont. Convex Optimization M2. 2/49 DMs DM par email: dm.daspremont@gmail.com A. d Aspremont. Convex Optimization

More information

MVE165/MMG631 Linear and integer optimization with applications Lecture 13 Overview of nonlinear programming. Ann-Brith Strömberg

MVE165/MMG631 Linear and integer optimization with applications Lecture 13 Overview of nonlinear programming. Ann-Brith Strömberg MVE165/MMG631 Overview of nonlinear programming Ann-Brith Strömberg 2015 05 21 Areas of applications, examples (Ch. 9.1) Structural optimization Design of aircraft, ships, bridges, etc Decide on the material

More information

Algorithms for constrained local optimization

Algorithms for constrained local optimization Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained

More information

(One Dimension) Problem: for a function f(x), find x 0 such that f(x 0 ) = 0. f(x)

(One Dimension) Problem: for a function f(x), find x 0 such that f(x 0 ) = 0. f(x) Solving Nonlinear Equations & Optimization One Dimension Problem: or a unction, ind 0 such that 0 = 0. 0 One Root: The Bisection Method This one s guaranteed to converge at least to a singularity, i not

More information

Scientific Computing: Optimization

Scientific Computing: Optimization Scientific Computing: Optimization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 March 8th, 2011 A. Donev (Courant Institute) Lecture

More information

Review of Classical Optimization

Review of Classical Optimization Part II Review of Classical Optimization Multidisciplinary Design Optimization of Aircrafts 51 2 Deterministic Methods 2.1 One-Dimensional Unconstrained Minimization 2.1.1 Motivation Most practical optimization

More information

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Lecture 15 Newton Method and Self-Concordance. October 23, 2008 Newton Method and Self-Concordance October 23, 2008 Outline Lecture 15 Self-concordance Notion Self-concordant Functions Operations Preserving Self-concordance Properties of Self-concordant Functions Implications

More information

CSCI 1951-G Optimization Methods in Finance Part 09: Interior Point Methods

CSCI 1951-G Optimization Methods in Finance Part 09: Interior Point Methods CSCI 1951-G Optimization Methods in Finance Part 09: Interior Point Methods March 23, 2018 1 / 35 This material is covered in S. Boyd, L. Vandenberge s book Convex Optimization https://web.stanford.edu/~boyd/cvxbook/.

More information

Numerical solutions of nonlinear systems of equations

Numerical solutions of nonlinear systems of equations Numerical solutions of nonlinear systems of equations Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan E-mail: min@math.ntnu.edu.tw August 28, 2011 Outline 1 Fixed points

More information

Optimization for Machine Learning

Optimization for Machine Learning Optimization for Machine Learning (Problems; Algorithms - A) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html

More information

arxiv: v3 [math.na] 23 Mar 2016

arxiv: v3 [math.na] 23 Mar 2016 Randomized Quasi-Newton Updates are Linearly Convergent Matrix Inversion Algorithms Robert M. Gower and Peter Richtárik arxiv:602.0768v3 [math.na] 23 Mar 206 School of Mathematics University of Edinburgh

More information

Nonlinear Optimization for Optimal Control

Nonlinear Optimization for Optimal Control Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]

More information

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection 6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE Three Alternatives/Remedies for Gradient Projection Two-Metric Projection Methods Manifold Suboptimization Methods

More information

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO QUESTION BOOKLET EECS 227A Fall 2009 Midterm Tuesday, Ocotober 20, 11:10-12:30pm DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO You have 80 minutes to complete the midterm. The midterm consists

More information

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Robert M. Freund March, 2004 2004 Massachusetts Institute of Technology. The Problem The logarithmic barrier approach

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Suppose that the approximate solutions of Eq. (1) satisfy the condition (3). Then (1) if η = 0 in the algorithm Trust Region, then lim inf.

Suppose that the approximate solutions of Eq. (1) satisfy the condition (3). Then (1) if η = 0 in the algorithm Trust Region, then lim inf. Maria Cameron 1. Trust Region Methods At every iteration the trust region methods generate a model m k (p), choose a trust region, and solve the constraint optimization problem of finding the minimum of

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted

More information

Optimization Methods

Optimization Methods Optimization Methods Categorization of Optimization Problems Continuous Optimization Discrete Optimization Combinatorial Optimization Variational Optimization Common Optimization Concepts in Computer Vision

More information

Chapter 8 Gradient Methods

Chapter 8 Gradient Methods Chapter 8 Gradient Methods An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Introduction Recall that a level set of a function is the set of points satisfying for some constant. Thus, a point

More information

Lecture 18: November Review on Primal-dual interior-poit methods

Lecture 18: November Review on Primal-dual interior-poit methods 10-725/36-725: Convex Optimization Fall 2016 Lecturer: Lecturer: Javier Pena Lecture 18: November 2 Scribes: Scribes: Yizhu Lin, Pan Liu Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

Lecture 3: Linesearch methods (continued). Steepest descent methods

Lecture 3: Linesearch methods (continued). Steepest descent methods Lecture 3: Linesearch methods (continued). Steepest descent methods Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 3: Linesearch methods (continued).

More information

ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, Please leave fractions as fractions, but simplify them, etc.

ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, Please leave fractions as fractions, but simplify them, etc. ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, 2015 1 Name: Solution Score: /100 This exam is closed-book. You must show ALL of your work for full credit. Please read the questions carefully.

More information

Optimization Methods. Lecture 19: Line Searches and Newton s Method

Optimization Methods. Lecture 19: Line Searches and Newton s Method 15.93 Optimization Methods Lecture 19: Line Searches and Newton s Method 1 Last Lecture Necessary Conditions for Optimality (identifies candidates) x local min f(x ) =, f(x ) PSD Slide 1 Sufficient Conditions

More information

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study International Journal of Mathematics And Its Applications Vol.2 No.4 (2014), pp.47-56. ISSN: 2347-1557(online) Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms:

More information

Quasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization

Quasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization Quasi-Newton Methods Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization 10-725 Last time: primal-dual interior-point methods Given the problem min x f(x) subject to h(x)

More information

Chapter 2. Optimization. Gradients, convexity, and ALS

Chapter 2. Optimization. Gradients, convexity, and ALS Chapter 2 Optimization Gradients, convexity, and ALS Contents Background Gradient descent Stochastic gradient descent Newton s method Alternating least squares KKT conditions 2 Motivation We can solve

More information

Optimality Conditions for Constrained Optimization

Optimality Conditions for Constrained Optimization 72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)

More information