Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem

Size: px
Start display at page:

Download "Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem"

Transcription

1 Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem min{f (x) : x R n }. The iterative algorithms that we will consider are of the form x k+1 = x k + t k d k, k = 0, 1,... d k - direction. t k - stepsize. We will limit ourselves to descent directions. Definition. Let f : R n R be a continuously differentiable function over R n. A vector 0 d R n is called a descent direction of f at x if the directional derivative f (x; d) is negative, meaning that f (x; d) = f (x) T d < 0. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 1 / 33

2 The Descent Property of Descent Directions Lemma: Let f be a continuously differentiable function over R n, and let x R n. Suppose that d is a descent direction of f at x. Then there exists ε > 0 such that f (x + td) < f (x) Proof. for any t (0, ε]. Since f (x; d) < 0, it follows from the definition of the directional derivative that f (x + td) f (x) lim = f (x; d) < 0. t 0 + t Therefore, ε > 0 such that f (x + td) f (x) < 0 t for any t (0, ε], which readily implies the desired result. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 2 / 33

3 Schematic Descent Direction Method Initialization: pick x 0 R n arbitrarily. General step: for any k = 0, 1, 2,... set (a) pick a descent direction d k. (b) find a stepsize t k satisfying f (x k + t k d k ) < f (x k ). (c) set x k+1 = x k + t k d k. (d) if a stopping criteria is satisfied, then STOP and x k+1 is the output. Of course, many details are missing in the above schematic algorithm: What is the starting point? How to choose the descent direction? What stepsize should be taken? What is the stopping criteria? Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 3 / 33

4 Stepsize Selection Rules constant stepsize - t k = t for any k. exact stepsize - t k is a minimizer of f along the ray x k + td k : t k argmin f (x k + td k ). t 0 backtracking 1 - The method requires three parameters: s > 0, α (0, 1), β (0, 1). Here we start with an initial stepsize t k = s. While f (x k ) f (x k + t k d k ) < αt k f (x k ) T d k. set t k := βt k Sufficient Decrease Property: f (x k ) f (x k + t k d k ) αt k f (x k ) T d k. 1 also referred to as Armijo Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 4 / 33

5 Exact Line Search for Quadratic Functions f (x) = x T Ax + 2b T x + c where A is an n n positive definite matrix, b R n and c R. Let x R n and let d R n be a descent direction of f at x. The objective is to find a solution to min f (x + td). t 0 In class Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 5 / 33

6 The Gradient Method - Taking the Direction of Minus the Gradient In the gradient method d k = f (x k ). This is a descent direction as long as f (x k ) 0 since f (x k ; f (x k )) = f (x k ) T f (x k ) = f (x k ) 2 < 0. In addition for being a descent direction, minus the gradient is also the steepest direction method. Lemma: Let f be a continuously differentiable function and let x R n be a non-stationary point ( f (x) 0). Then an optimal solution of is d = f (x)/ f (x). Proof. In class min d {f (x; d) : d = 1} (1) Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 6 / 33

7 The Gradient Method The Gradient Method Input: ε > 0 - tolerance parameter. Initialization: pick x 0 R n arbitrarily. General step: for any k = 0, 1, 2,... execute the following steps: (a) pick a stepsize t k by a line search procedure on the function g(t) = f (x k t f (x k )). (b) set x k+1 = x k t k f (x k ). (c) if f (x k+1 ) ε, then STOP and x k+1 is the output. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 7 / 33

8 Numerical Example x 0 = (2; 1), ε = 10 5, exact line search. min x 2 + 2y 2 13 iterations until convergence. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 8 / 33

9 The Zig-Zag Effect Lemma. Let {x k } k 0 be the sequence generated by the gradient method with exact line search for solving a problem of minimizing a continuously differentiable function f. Then for any k = 0, 1, 2,... (x k+2 x k+1 ) T (x k+1 x k ) = 0. Proof. x k+1 x k = t k f (x k ), x k+2 x k+1 = t k+1 f (x k+1 ). Therefore, we need to prove that f (x k ) T f (x k+1 ) = 0. t k argmin{g(t) f (x k t f (x k ))} t 0 Hence, g (t k ) = 0. f (x k ) T f (x k t k f (x k )) = 0. f (x k ) T f (x k+1 ) = 0. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 9 / 33

10 Numerical Example - Constant Stepsize, t = 0.1 x 0 = (2; 1), ε = 10 5, t = 0.1. min x 2 + 2y 2 iter_number = 1 norm_grad = fun_val = iter_number = 2 norm_grad = fun_val = iter_number = 3 norm_grad = fun_val = : : : iter_number = 56 norm_grad = fun_val = iter_number = 57 norm_grad = fun_val = iter_number = 58 norm_grad = fun_val = quite a lot of iterations... Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 10 / 33

11 Numerical Example - Constant Stepsize, t = 10 x 0 = (2; 1), ε = 10 5, t = 10.. min x 2 + 2y 2 iter_number = 1 norm_grad = fun_val = iter_number = 2 norm_grad = fun_val = iter_number = 3 norm_grad = fun_val = : : : iter_number = 119 norm_grad = NaN fun_val = NaN The sequence diverges:( Important question: how can we choose the constant stepsize so that convergence is guaranteed? Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 11 / 33

12 Lipschitz Continuity of the Gradient Definition Let f be a continuously differentiable function over R n. We say that f has a Lipschitz gradient if there exists L 0 for which f (x) f (y) L x y for any x, y R n. L is called the Lipschitz constant. If f is Lipschitz with constant L, then it is also Lipschitz with constant L for all L L. The class of functions with Lipschitz gradient with constant L is denoted by C 1,1 L (Rn ) or just C 1,1 L. Linear functions - Given a R n, the function f (x) = a T x is in C 1,1 0. Quadratic functions - Let A be a symmetric n n matrix, b R n and c R. Then the function f (x) = x T Ax + 2b T x + c is a C 1,1 function. The smallest Lipschitz constant of f is A 2 why? In class Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 12 / 33

13 Equivalence to Boundedness of the Hessian Theorem. Let f be a twice continuously differentiable function over R n. Then the following two claims are equivalent: 1. f C 1,1 L (Rn ) f (x) L for any x R n. Proof on pages 73,74 of the book Example: f (x) = 1 + x 2 C 1,1 In class Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 13 / 33

14 Convergence of the Gradient Method Theorem. Let {x k } k 0 be the sequence generated by GM for solving min f (x) x R n with one of the following stepsize strategies: constant stepsize t ( 0, 2 L). exact line search. backtracking procedure with parameters s > 0 and α, β (0, 1). Assume that f C 1,1 L (Rn ). f is bounded below over R n, that is, there exists m R such that f (x) > m for all x R n ). Then 1. for any k, f (x k+1 ) < f (x k ) unless f (x k ) = f (x k ) 0 as k. Theorem 4.25 in the book. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 14 / 33

15 Two Numerical Examples - Backtracking min x 2 + 2y 2 x 0 = (2; 1), s = 2, α = 0.25, β = 0.5, ε = iter_number = 1 norm_grad = fun_val = iter_number = 2 norm_grad = fun_val = fast convergence (also due to lack!) no real advantage to exact line search. ANOTHER EXAMPLE: min 0.01x 2 + y 2, s = 2, α = 0.25, β = 0.5, ε = iter_number = 1 norm_grad = fun_val = iter_number = 2 norm_grad = fun_val = iter_number = 3 norm_grad = fun_val = : : : iter_number = 201 norm_grad = fun_val = Important Question: Can we detect key properties of the objective function that imply slow/fast convergence? Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 15 / 33

16 Kantorovich Inequality Lemma. Let A be a positive definite n n matrix. Then for any 0 x R n the inequality Proof. holds. x T x (x T Ax)(x T A 1 x) 4λ max(a)λ min (A) (λ max (A) + λ min (A)) 2 Denote m = λ min (A) and M = λ max (A). The eigenvalues of the matrix A + MmA 1 are λ i (A) + Mm λ i (A). The maximum of the 1-D function ϕ(t) = t + Mm t over [m, M] is attained at the endpoints m and M with a corresponding value of M + m. Thus, the eigenvalues of A + MmA 1 are smaller than (M + m). A + MmA 1 (M + m)i. x T Ax + Mm(x T A 1 x) (M + m)(x T x), Therefore, (x T Ax)[Mm(x T A 1 x)] 1 [ (x T Ax) + Mm(x T A 1 x) ] 2 (M + m) 2 (x T x) 2, 4 4 Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 16 / 33

17 Gradient Method for Minimizing x T Ax Theorem. Let {x k } k 0 be the sequence generated by the gradient method with exact linesearch for solving the problem Then for any k = 0, 1,...: min x R xt Ax (A 0). n f (x k+1 ) where M = λ max (A), m = λ min (A). ( ) 2 M m f (x k ), M + m Proof. x k+1 = x k t k d k, where t k = dt k d k 2d T k Ad, d k = 2Ax k. k Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 17 / 33

18 Proof of Rate of Convergence Contd. f (x k+1 ) = x T k+1ax k+1 = (x k t k d k ) T A(x k t k d k ) = x T k Ax k 2t k d T k Ax k + tk 2 d T k Ad k = x T k Ax k t k d T k d k + tk 2 d T k Ad k. Plugging in the expression for t k By Kantorovich: f (x k+1 ) f (x k+1 ) = x T k Ax k 1 (d T k d k ) 2 4 d T k Ad k ( = x T k Ax k = ( 1 4Mm (M + m) 2 ( 1 ) (d T k d k ) 2 (d T k Ad k)(x T k AA 1 Ax k ) ) f (x k ). (d T k d k ) 2 (d T k Ad k)(d T k A 1 d k ) ) f (x k ) = ( ) 2 M m f (x k ) = M + m ( ) 2 κ(a) 1 f (x k ), κ(a) + 1 Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 18 / 33

19 The Condition Number Definition. Let A be an n n positive definite matrix. Then the condition number of A is defined by κ(a) = λ max(a) λ min (A). matrices (or quadratic functions) with large condition number are called ill-conditioned. matrices with small condition number are called well-conditioned. large condition number implies large number of iterations of the gradient method. small condition number implies small number of iterations of the gradient method. For a non-quadratic function, the asymptotic rate of convergence of x k to a stationary point x is usually determined by the condition number of 2 f (x ). Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 19 / 33

20 A Severely Ill-Condition Function - Rosenbrock min { f (x 1, x 2 ) = 100(x 2 x 2 1 ) 2 + (1 x 1 ) 2}. optimal solution:(x 1, x 2 ) = (1, 1), optimal value: 0. f (x) = 2 f (x) = ( 400x1 (x 2 x1 2) 2(1 x ) 1) 200(x 2 x1 2), ( 400x x x ) x condition number: f (1, 1) = ( 802 ) Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 20 / 33

21 Solution of the Rosenbrock Problem with the Gradient Method x 0 = (2; 5), s = 2, α = 0.25, β = 0.5, ε = 10 5, backtracking stepsize selection. 6890(!!!) iterations. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 21 / 33

22 Sensitivity of Solutions to Linear Systems Suppose that we are given the linear system Ax = b where A 0 and we assume that x is indeed the solution of the system (x = A 1 b). Suppose that the right-hand side is perturbed b + b. What can be said on the solution of the new system x + x? x = A 1 b. Result (derivation In class): x x κ(a) b b Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 22 / 33

23 Numerical Example consider the ill-condition matrix: ( ) A = >> A=[1+1e-5,1;1,1+1e-5]; >> cond(a) ans = e+005 We have >> A\[1;1] ans = However, >> A\[1.1;1] ans = 1.0e+003 * Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 23 / 33

24 x k+1 = x k t k D f (x k ). Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 24 / 33 Scaled Gradient Method Consider the minimization problem (P) min{f (x) : x R n }. For a given nonsingular matrix S R n n, we make the linear change of variables x = Sy, and obtain the equivalent problem (P ) min{g(y) f (Sy) : y R n }. Since g(y) = S T f (Sy) = S T f (x), the gradient method for (P ) is y k+1 = y k t k S T f (Sy k ). Multiplying the latter equality by S from the left, and using the notation x k = Sy k : x k+1 = x k t k SS T f (x k ). Defining D = SS T, we obtain the scaled gradient method:

25 Scaled Gradient Method D 0, so the direction D f (x k ) is a descent direction: f (x k ; D f (x k )) = f (x k ) T D f (x k ) < 0, We also allow different scaling matrices at each iteration. Scaled Gradient Method Input: ε > 0 - tolerance parameter. Initialization: pick x 0 R n arbitrarily. General step: for any k = 0, 1, 2,... execute the following steps: (a) pick a scaling matrix D k 0. (b) pick a stepsize t k by a line search procedure on the function (c) set x k+1 = x k t k D k f (x k ). g(t) = f (x k td k f (x k )). (c) if f (x k+1 ) ε, then STOP and x k+1 is the output. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 25 / 33

26 Choosing the Scaling Matrix D k The scaled gradient method with scaling matrix D is equivalent to the gradient method employed on the function g(y) = f (D 1/2 y). Note that the gradient and Hessian of g are given by g(y) = D 1/2 f (D 1/2 y) = D 1/2 f (x), 2 g(y) = D 1/2 2 f (D 1/2 y)d 1/2 = D 1/2 2 f (x)d 1/2.. The objective is usually to pick D k so as to make D 1/2 k 2 f (x k )D 1/2 k as well-conditioned as possible. A well known choice (Newton s method): D k = ( 2 f (x k )) 1. diagonal scaling: D k is picked to be diagonal. For example, ( 2 ) 1 f (x k ) (D k ) ii =. Diagonal scaling can be very effective when the decision variables are of different magnitudes. x 2 i Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 26 / 33

27 The Gauss-Newton Method Nonlinear least squares problem: { (NLS): min x R n g(x) } m (f i (x) c i ) 2. f 1,..., f m are continuously differentiable over R n and c 1,..., c m R. Denote: Then the problem becomes: i=1 f 1 (x) c 1 f 2 (x) c 2 F (x) =., f m (x) c m min F (x) 2. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 27 / 33

28 The Gauss-Newton Method Given the kth iterate x k, the next iterate is chosen to minimize the sum of squares of the linearized terms, that is, { m } [ x k+1 = argmin fi (x k ) + f i (x k ) T ] 2 (x x k ) c i. x R n i=1 The general step actually consists of solving the linear LS problem min A k x b k 2, where f 1(x k ) T f 2(x k ) T A k =. = J(x k) f m(x k ) T is the so-called Jacobian matrix, assumed to have full column rank. f 1(x k ) T x k f 1(x k ) + c 1 f 2(x k ) T x k f 2(x k ) + c 2 b k =. = J(x k)x k F (x k ) f m(x k ) T x k f m(x k ) + c m Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 28 / 33

29 The Gauss-Newton Method The Gauss-Newton method can thus be written as: x k+1 = (J(x k ) T J(x k )) 1 J(x k ) T b k. The gradient of the objective function f (x) = F (x) 2 is f (x) = 2J(x)F (x) The GN method can be rewritten as follows: x k+1 = (J(x k ) T J(x k )) 1 J(x k ) T (J(x k )x k F (x k )) = x k (J(x k ) T J(x k )) 1 J(x k ) T F (x k ) = x k 1 2 (J(x k) T J(x k )) 1 f (x k ), that is, it is a scaled gradient method with a special choice of scaling matrix: D k = 1 2 (J(x k) T J(x k )) 1. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 29 / 33

30 The Damped Gauss-Newton Method The Gauss-Newton method does not incorporate a stepsize, which might cause it to diverge. A well known variation of the method incorporating stepsizes is the damped Gauss-newton Method. Damped Gauss-Newton Method Input: ε - tolerance parameter. Initialization: pick x 0 R n arbitrarily. General step: for any k = 0, 1, 2,... execute the following steps: (a) Set d k = (J(x k ) T J(x k )) 1 J(x k )F (x k ). (b) Set t k by a line search procedure on the function (c) set x k+1 = x k + t k d k. h(t) = g(x k + td k ). (c) if f (x k+1 ) ε, then STOP and x k+1 is the output. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 30 / 33

31 Fermat-Weber Problem Fermat-Weber Problem: Given m points in R n : a 1,..., a m also called anchor point and m weights ω 1, ω 2,..., ω m > 0, find a point x R n that minimizes the weighted distance of x to each of the points a 1,..., a m : { } min x R n f (x) m ω i x a i i=1. The objective function is not differentiable at the anchor points a 1,..., a m. One of the simplest instances of facility location problems. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 31 / 33

32 Weiszfeld s Method (1937) Start from the stationarity condition f (x) = 0. 2 m i=1 ω x a i i ( m ω i i=1 x a i x = 1 x a i = 0. ) x = m i=1 m ω i a i m ω i i=1 x a i=1 x a i i. ω i a i x a i, The stationarity condition can be written as x = T (x), where T is the operator 1 m ω i a i T (x) x a i. m i=1 ω i x a i Weiszfeld s method is a fixed point method: i=1 x k+1 = T (x k ). 2 We implicitly assume here that x is not an anchor point. Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 32 / 33

33 Weiszfeld s Method as a Gradient Method Weiszfeld s Method Initialization: pick x 0 R n such that x a 1, a 2,..., a m. General step: for any k = 0, 1, 2,... compute: x k+1 = T (x k ) = m i=1 1 ω i x k a i m i=1 ω i a i x k a i. Weiszfeld s method is a gradient method since 1 m x k+1 = m i=1 = x k = x k ω i x k a i m i=1 m i=1 1 i=1 ω i x k a i 1 ω i x k a i ω i a i x k a i m i=1 ω i f (x k ). x k a i x k a i A gradient method with a special choice of stepsize: t k = 1 m i=1 ω i x k a i Amir Beck Introduction to Nonlinear Optimization Lecture Slides - The Gradient Method 33 / 33.

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem min{f (x) : x R n }. The iterative algorithms that we will consider are of the form x k+1 = x k + t k d k, k = 0, 1,...

More information

Unconstrained minimization of smooth functions

Unconstrained minimization of smooth functions Unconstrained minimization of smooth functions We want to solve min x R N f(x), where f is convex. In this section, we will assume that f is differentiable (so its gradient exists at every point), and

More information

Unconstrained optimization

Unconstrained optimization Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout

More information

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then 1. x S is a global minimum point of f over S if f (x) f (x ) for any x S. 2. x S

More information

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term;

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term; Chapter 2 Gradient Methods The gradient method forms the foundation of all of the schemes studied in this book. We will provide several complementary perspectives on this algorithm that highlight the many

More information

The Steepest Descent Algorithm for Unconstrained Optimization

The Steepest Descent Algorithm for Unconstrained Optimization The Steepest Descent Algorithm for Unconstrained Optimization Robert M. Freund February, 2014 c 2014 Massachusetts Institute of Technology. All rights reserved. 1 1 Steepest Descent Algorithm The problem

More information

Optimization Methods. Lecture 19: Line Searches and Newton s Method

Optimization Methods. Lecture 19: Line Searches and Newton s Method 15.93 Optimization Methods Lecture 19: Line Searches and Newton s Method 1 Last Lecture Necessary Conditions for Optimality (identifies candidates) x local min f(x ) =, f(x ) PSD Slide 1 Sufficient Conditions

More information

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Lecture 15 Newton Method and Self-Concordance. October 23, 2008 Newton Method and Self-Concordance October 23, 2008 Outline Lecture 15 Self-concordance Notion Self-concordant Functions Operations Preserving Self-concordance Properties of Self-concordant Functions Implications

More information

5 Overview of algorithms for unconstrained optimization

5 Overview of algorithms for unconstrained optimization IOE 59: NLP, Winter 22 c Marina A. Epelman 9 5 Overview of algorithms for unconstrained optimization 5. General optimization algorithm Recall: we are attempting to solve the problem (P) min f(x) s.t. x

More information

Lecture 3: Linesearch methods (continued). Steepest descent methods

Lecture 3: Linesearch methods (continued). Steepest descent methods Lecture 3: Linesearch methods (continued). Steepest descent methods Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 3: Linesearch methods (continued).

More information

Line Search Methods for Unconstrained Optimisation

Line Search Methods for Unconstrained Optimisation Line Search Methods for Unconstrained Optimisation Lecture 8, Numerical Linear Algebra and Optimisation Oxford University Computing Laboratory, MT 2007 Dr Raphael Hauser (hauser@comlab.ox.ac.uk) The Generic

More information

Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

More information

10. Unconstrained minimization

10. Unconstrained minimization Convex Optimization Boyd & Vandenberghe 10. Unconstrained minimization terminology and assumptions gradient descent method steepest descent method Newton s method self-concordant functions implementation

More information

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Coralia Cartis, University of Oxford INFOMM CDT: Modelling, Analysis and Computation of Continuous Real-World Problems Methods

More information

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization E5295/5B5749 Convex optimization with engineering applications Lecture 8 Smooth convex unconstrained and equality-constrained minimization A. Forsgren, KTH 1 Lecture 8 Convex optimization 2006/2007 Unconstrained

More information

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are Quadratic forms We consider the quadratic function f : R 2 R defined by f(x) = 2 xt Ax b T x with x = (x, x 2 ) T, () where A R 2 2 is symmetric and b R 2. We will see that, depending on the eigenvalues

More information

Written Examination

Written Examination Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes

More information

14. Nonlinear equations

14. Nonlinear equations L. Vandenberghe ECE133A (Winter 2018) 14. Nonlinear equations Newton method for nonlinear equations damped Newton method for unconstrained minimization Newton method for nonlinear least squares 14-1 Set

More information

4 damped (modified) Newton methods

4 damped (modified) Newton methods 4 damped (modified) Newton methods 4.1 damped Newton method Exercise 4.1 Determine with the damped Newton method the unique real zero x of the real valued function of one variable f(x) = x 3 +x 2 using

More information

Unconstrained Optimization

Unconstrained Optimization 1 / 36 Unconstrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University February 2, 2015 2 / 36 3 / 36 4 / 36 5 / 36 1. preliminaries 1.1 local approximation

More information

Conjugate Gradient (CG) Method

Conjugate Gradient (CG) Method Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous

More information

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44 Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)

More information

Unconstrained minimization: assumptions

Unconstrained minimization: assumptions Unconstrained minimization I terminology and assumptions I gradient descent method I steepest descent method I Newton s method I self-concordant functions I implementation IOE 611: Nonlinear Programming,

More information

Gradient Descent. Ryan Tibshirani Convex Optimization /36-725

Gradient Descent. Ryan Tibshirani Convex Optimization /36-725 Gradient Descent Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: canonical convex programs Linear program (LP): takes the form min x subject to c T x Gx h Ax = b Quadratic program (QP): like

More information

Descent methods. min x. f(x)

Descent methods. min x. f(x) Gradient Descent Descent methods min x f(x) 5 / 34 Descent methods min x f(x) x k x k+1... x f(x ) = 0 5 / 34 Gradient methods Unconstrained optimization min f(x) x R n. 6 / 34 Gradient methods Unconstrained

More information

Math 273a: Optimization Netwon s methods

Math 273a: Optimization Netwon s methods Math 273a: Optimization Netwon s methods Instructor: Wotao Yin Department of Mathematics, UCLA Fall 2015 some material taken from Chong-Zak, 4th Ed. Main features of Newton s method Uses both first derivatives

More information

Lecture 5: Gradient Descent. 5.1 Unconstrained minimization problems and Gradient descent

Lecture 5: Gradient Descent. 5.1 Unconstrained minimization problems and Gradient descent 10-725/36-725: Convex Optimization Spring 2015 Lecturer: Ryan Tibshirani Lecture 5: Gradient Descent Scribes: Loc Do,2,3 Disclaimer: These notes have not been subjected to the usual scrutiny reserved for

More information

Constrained Optimization

Constrained Optimization 1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange

More information

Numerical Optimization Prof. Shirish K. Shevade Department of Computer Science and Automation Indian Institute of Science, Bangalore

Numerical Optimization Prof. Shirish K. Shevade Department of Computer Science and Automation Indian Institute of Science, Bangalore Numerical Optimization Prof. Shirish K. Shevade Department of Computer Science and Automation Indian Institute of Science, Bangalore Lecture - 13 Steepest Descent Method Hello, welcome back to this series

More information

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization 5.93 Optimization Methods Lecture 8: Optimality Conditions and Gradient Methods for Unconstrained Optimization Outline. Necessary and sucient optimality conditions Slide. Gradient m e t h o d s 3. The

More information

AM 205: lecture 18. Last time: optimization methods Today: conditions for optimality

AM 205: lecture 18. Last time: optimization methods Today: conditions for optimality AM 205: lecture 18 Last time: optimization methods Today: conditions for optimality Existence of Global Minimum For example: f (x, y) = x 2 + y 2 is coercive on R 2 (global min. at (0, 0)) f (x) = x 3

More information

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method. Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization

More information

Unconstrained minimization

Unconstrained minimization CSCI5254: Convex Optimization & Its Applications Unconstrained minimization terminology and assumptions gradient descent method steepest descent method Newton s method self-concordant functions 1 Unconstrained

More information

Self-Concordant Barrier Functions for Convex Optimization

Self-Concordant Barrier Functions for Convex Optimization Appendix F Self-Concordant Barrier Functions for Convex Optimization F.1 Introduction In this Appendix we present a framework for developing polynomial-time algorithms for the solution of convex optimization

More information

A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable)

A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable) Nonlinear equations Definition A nonlinear equation is any equation of the form where f is a nonlinear function. Nonlinear equations x 2 + x + 1 = 0 (f : R R) f(x) = 0 (x cos y, 2y sin x) = (0, 0) (f :

More information

Nonlinear Optimization

Nonlinear Optimization Nonlinear Optimization (Com S 477/577 Notes) Yan-Bin Jia Nov 7, 2017 1 Introduction Given a single function f that depends on one or more independent variable, we want to find the values of those variables

More information

Lecture Notes: Geometric Considerations in Unconstrained Optimization

Lecture Notes: Geometric Considerations in Unconstrained Optimization Lecture Notes: Geometric Considerations in Unconstrained Optimization James T. Allison February 15, 2006 The primary objectives of this lecture on unconstrained optimization are to: Establish connections

More information

Numerical Optimization of Partial Differential Equations

Numerical Optimization of Partial Differential Equations Numerical Optimization of Partial Differential Equations Part I: basic optimization concepts in R n Bartosz Protas Department of Mathematics & Statistics McMaster University, Hamilton, Ontario, Canada

More information

17 Solution of Nonlinear Systems

17 Solution of Nonlinear Systems 17 Solution of Nonlinear Systems We now discuss the solution of systems of nonlinear equations. An important ingredient will be the multivariate Taylor theorem. Theorem 17.1 Let D = {x 1, x 2,..., x m

More information

Nonlinear Optimization for Optimal Control

Nonlinear Optimization for Optimal Control Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]

More information

Lecture 5: September 15

Lecture 5: September 15 10-725/36-725: Convex Optimization Fall 2015 Lecture 5: September 15 Lecturer: Lecturer: Ryan Tibshirani Scribes: Scribes: Di Jin, Mengdi Wang, Bin Deng Note: LaTeX template courtesy of UC Berkeley EECS

More information

Introduction to Nonlinear Optimization Paul J. Atzberger

Introduction to Nonlinear Optimization Paul J. Atzberger Introduction to Nonlinear Optimization Paul J. Atzberger Comments should be sent to: atzberg@math.ucsb.edu Introduction We shall discuss in these notes a brief introduction to nonlinear optimization concepts,

More information

Convex Optimization. 9. Unconstrained minimization. Prof. Ying Cui. Department of Electrical Engineering Shanghai Jiao Tong University

Convex Optimization. 9. Unconstrained minimization. Prof. Ying Cui. Department of Electrical Engineering Shanghai Jiao Tong University Convex Optimization 9. Unconstrained minimization Prof. Ying Cui Department of Electrical Engineering Shanghai Jiao Tong University 2017 Autumn Semester SJTU Ying Cui 1 / 40 Outline Unconstrained minimization

More information

8 Numerical methods for unconstrained problems

8 Numerical methods for unconstrained problems 8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields

More information

Lecture 5: September 12

Lecture 5: September 12 10-725/36-725: Convex Optimization Fall 2015 Lecture 5: September 12 Lecturer: Lecturer: Ryan Tibshirani Scribes: Scribes: Barun Patra and Tyler Vuong Note: LaTeX template courtesy of UC Berkeley EECS

More information

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained

More information

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares Robert Bridson October 29, 2008 1 Hessian Problems in Newton Last time we fixed one of plain Newton s problems by introducing line search

More information

Newton s Method. Javier Peña Convex Optimization /36-725

Newton s Method. Javier Peña Convex Optimization /36-725 Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and

More information

c 2007 Society for Industrial and Applied Mathematics

c 2007 Society for Industrial and Applied Mathematics SIAM J. OPTIM. Vol. 18, No. 1, pp. 106 13 c 007 Society for Industrial and Applied Mathematics APPROXIMATE GAUSS NEWTON METHODS FOR NONLINEAR LEAST SQUARES PROBLEMS S. GRATTON, A. S. LAWLESS, AND N. K.

More information

March 8, 2010 MATH 408 FINAL EXAM SAMPLE

March 8, 2010 MATH 408 FINAL EXAM SAMPLE March 8, 200 MATH 408 FINAL EXAM SAMPLE EXAM OUTLINE The final exam for this course takes place in the regular course classroom (MEB 238) on Monday, March 2, 8:30-0:20 am. You may bring two-sided 8 page

More information

Mathematical modelling Chapter 2 Nonlinear models and geometric models

Mathematical modelling Chapter 2 Nonlinear models and geometric models Mathematical modelling Chapter 2 Nonlinear models and geometric models Faculty of Computer and Information Science University of Ljubljana 2018/2019 3. Nonlinear models Given is a sample of points {(x

More information

6. Proximal gradient method

6. Proximal gradient method L. Vandenberghe EE236C (Spring 2016) 6. Proximal gradient method motivation proximal mapping proximal gradient method with fixed step size proximal gradient method with line search 6-1 Proximal mapping

More information

Optimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29

Optimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29 Optimization Yuh-Jye Lee Data Science and Machine Intelligence Lab National Chiao Tung University March 21, 2017 1 / 29 You Have Learned (Unconstrained) Optimization in Your High School Let f (x) = ax

More information

Chapter 0. Mathematical Preliminaries. 0.1 Norms

Chapter 0. Mathematical Preliminaries. 0.1 Norms Chapter 0 Mathematical Preliminaries 0.1 Norms Throughout this course we will be working with the vector space R n. For this reason we begin with a brief review of its metric space properties Definition

More information

Higher-Order Methods

Higher-Order Methods Higher-Order Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. PCMI, July 2016 Stephen Wright (UW-Madison) Higher-Order Methods PCMI, July 2016 1 / 25 Smooth

More information

OPER 627: Nonlinear Optimization Lecture 2: Math Background and Optimality Conditions

OPER 627: Nonlinear Optimization Lecture 2: Math Background and Optimality Conditions OPER 627: Nonlinear Optimization Lecture 2: Math Background and Optimality Conditions Department of Statistical Sciences and Operations Research Virginia Commonwealth University Aug 28, 2013 (Lecture 2)

More information

Complexity of gradient descent for multiobjective optimization

Complexity of gradient descent for multiobjective optimization Complexity of gradient descent for multiobjective optimization J. Fliege A. I. F. Vaz L. N. Vicente July 18, 2018 Abstract A number of first-order methods have been proposed for smooth multiobjective optimization

More information

Part 2: Linesearch methods for unconstrained optimization. Nick Gould (RAL)

Part 2: Linesearch methods for unconstrained optimization. Nick Gould (RAL) Part 2: Linesearch methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428

More information

Nonlinear equations. Norms for R n. Convergence orders for iterative methods

Nonlinear equations. Norms for R n. Convergence orders for iterative methods Nonlinear equations Norms for R n Assume that X is a vector space. A norm is a mapping X R with x such that for all x, y X, α R x = = x = αx = α x x + y x + y We define the following norms on the vector

More information

Image restoration: numerical optimisation

Image restoration: numerical optimisation Image restoration: numerical optimisation Short and partial presentation Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP / 6 Context

More information

March 5, 2012 MATH 408 FINAL EXAM SAMPLE

March 5, 2012 MATH 408 FINAL EXAM SAMPLE March 5, 202 MATH 408 FINAL EXAM SAMPLE Partial Solutions to Sample Questions (in progress) See the sample questions for the midterm exam, but also consider the following questions. Obviously, a final

More information

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method L. Vandenberghe EE236C (Spring 2016) 1. Gradient method gradient method, first-order methods quadratic bounds on convex functions analysis of gradient method 1-1 Approximate course outline First-order

More information

6. Proximal gradient method

6. Proximal gradient method L. Vandenberghe EE236C (Spring 2013-14) 6. Proximal gradient method motivation proximal mapping proximal gradient method with fixed step size proximal gradient method with line search 6-1 Proximal mapping

More information

Unconstrained optimization I Gradient-type methods

Unconstrained optimization I Gradient-type methods Unconstrained optimization I Gradient-type methods Antonio Frangioni Department of Computer Science University of Pisa www.di.unipi.it/~frangio frangio@di.unipi.it Computational Mathematics for Learning

More information

TMA4180 Solutions to recommended exercises in Chapter 3 of N&W

TMA4180 Solutions to recommended exercises in Chapter 3 of N&W TMA480 Solutions to recommended exercises in Chapter 3 of N&W Exercise 3. The steepest descent and Newtons method with the bactracing algorithm is implemented in rosenbroc_newton.m. With initial point

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Some definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts

Some definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts Some definitions Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization M. M. Sussman sussmanm@math.pitt.edu Office Hours: MW 1:45PM-2:45PM, Thack 622 A matrix A is SPD (Symmetric

More information

CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS

CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS Igor V. Konnov Department of Applied Mathematics, Kazan University Kazan 420008, Russia Preprint, March 2002 ISBN 951-42-6687-0 AMS classification:

More information

Lecture 14: October 17

Lecture 14: October 17 1-725/36-725: Convex Optimization Fall 218 Lecture 14: October 17 Lecturer: Lecturer: Ryan Tibshirani Scribes: Pengsheng Guo, Xian Zhou Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

Conjugate Gradient Method

Conjugate Gradient Method Conjugate Gradient Method Hung M Phan UMass Lowell April 13, 2017 Throughout, A R n n is symmetric and positive definite, and b R n 1 Steepest Descent Method We present the steepest descent method for

More information

10.34 Numerical Methods Applied to Chemical Engineering Fall Quiz #1 Review

10.34 Numerical Methods Applied to Chemical Engineering Fall Quiz #1 Review 10.34 Numerical Methods Applied to Chemical Engineering Fall 2015 Quiz #1 Review Study guide based on notes developed by J.A. Paulson, modified by K. Severson Linear Algebra We ve covered three major topics

More information

Network Newton. Aryan Mokhtari, Qing Ling and Alejandro Ribeiro. University of Pennsylvania, University of Science and Technology (China)

Network Newton. Aryan Mokhtari, Qing Ling and Alejandro Ribeiro. University of Pennsylvania, University of Science and Technology (China) Network Newton Aryan Mokhtari, Qing Ling and Alejandro Ribeiro University of Pennsylvania, University of Science and Technology (China) aryanm@seas.upenn.edu, qingling@mail.ustc.edu.cn, aribeiro@seas.upenn.edu

More information

Tangent spaces, normals and extrema

Tangent spaces, normals and extrema Chapter 3 Tangent spaces, normals and extrema If S is a surface in 3-space, with a point a S where S looks smooth, i.e., without any fold or cusp or self-crossing, we can intuitively define the tangent

More information

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL) Part 3: Trust-region methods for unconstrained optimization Nick Gould (RAL) minimize x IR n f(x) MSc course on nonlinear optimization UNCONSTRAINED MINIMIZATION minimize x IR n f(x) where the objective

More information

Mathematical optimization

Mathematical optimization Optimization Mathematical optimization Determine the best solutions to certain mathematically defined problems that are under constrained determine optimality criteria determine the convergence of the

More information

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for 1 Iteration basics Notes for 2016-11-07 An iterative solver for Ax = b is produces a sequence of approximations x (k) x. We always stop after finitely many steps, based on some convergence criterion, e.g.

More information

CS 450 Numerical Analysis. Chapter 5: Nonlinear Equations

CS 450 Numerical Analysis. Chapter 5: Nonlinear Equations Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80

More information

Optimization Methods

Optimization Methods Optimization Methods Decision making Examples: determining which ingredients and in what quantities to add to a mixture being made so that it will meet specifications on its composition allocating available

More information

Numerical solutions of nonlinear systems of equations

Numerical solutions of nonlinear systems of equations Numerical solutions of nonlinear systems of equations Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan E-mail: min@math.ntnu.edu.tw August 28, 2011 Outline 1 Fixed points

More information

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen Numerisches Rechnen (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang Institut für Geometrie und Praktische Mathematik RWTH Aachen Wintersemester 2011/12 IGPM, RWTH Aachen Numerisches Rechnen

More information

Miscellaneous Nonlinear Programming Exercises

Miscellaneous Nonlinear Programming Exercises Miscellaneous Nonlinear Programming Exercises Henry Wolkowicz 2 08 21 University of Waterloo Department of Combinatorics & Optimization Waterloo, Ontario N2L 3G1, Canada Contents 1 Numerical Analysis Background

More information

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization Proximal Newton Method Zico Kolter (notes by Ryan Tibshirani) Convex Optimization 10-725 Consider the problem Last time: quasi-newton methods min x f(x) with f convex, twice differentiable, dom(f) = R

More information

C&O367: Nonlinear Optimization (Winter 2013) Assignment 4 H. Wolkowicz

C&O367: Nonlinear Optimization (Winter 2013) Assignment 4 H. Wolkowicz C&O367: Nonlinear Optimization (Winter 013) Assignment 4 H. Wolkowicz Posted Mon, Feb. 8 Due: Thursday, Feb. 8 10:00AM (before class), 1 Matrices 1.1 Positive Definite Matrices 1. Let A S n, i.e., let

More information

IE 5531: Engineering Optimization I

IE 5531: Engineering Optimization I IE 5531: Engineering Optimization I Lecture 15: Nonlinear optimization Prof. John Gunnar Carlsson November 1, 2010 Prof. John Gunnar Carlsson IE 5531: Engineering Optimization I November 1, 2010 1 / 24

More information

Arc Search Algorithms

Arc Search Algorithms Arc Search Algorithms Nick Henderson and Walter Murray Stanford University Institute for Computational and Mathematical Engineering November 10, 2011 Unconstrained Optimization minimize x D F (x) where

More information

Scientific Computing: Optimization

Scientific Computing: Optimization Scientific Computing: Optimization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 March 8th, 2011 A. Donev (Courant Institute) Lecture

More information

On the convergence properties of the modified Polak Ribiére Polyak method with the standard Armijo line search

On the convergence properties of the modified Polak Ribiére Polyak method with the standard Armijo line search ANZIAM J. 55 (E) pp.e79 E89, 2014 E79 On the convergence properties of the modified Polak Ribiére Polyak method with the standard Armijo line search Lijun Li 1 Weijun Zhou 2 (Received 21 May 2013; revised

More information

Optimization Tutorial 1. Basic Gradient Descent

Optimization Tutorial 1. Basic Gradient Descent E0 270 Machine Learning Jan 16, 2015 Optimization Tutorial 1 Basic Gradient Descent Lecture by Harikrishna Narasimhan Note: This tutorial shall assume background in elementary calculus and linear algebra.

More information

Gradient Descent. Lecturer: Pradeep Ravikumar Co-instructor: Aarti Singh. Convex Optimization /36-725

Gradient Descent. Lecturer: Pradeep Ravikumar Co-instructor: Aarti Singh. Convex Optimization /36-725 Gradient Descent Lecturer: Pradeep Ravikumar Co-instructor: Aarti Singh Convex Optimization 10-725/36-725 Based on slides from Vandenberghe, Tibshirani Gradient Descent Consider unconstrained, smooth convex

More information

Numerical optimization

Numerical optimization Numerical optimization Lecture 4 Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Numerical geometry of non-rigid shapes Stanford University, Winter 2009 2 Longest Slowest Shortest Minimal Maximal

More information

A globally and quadratically convergent primal dual augmented Lagrangian algorithm for equality constrained optimization

A globally and quadratically convergent primal dual augmented Lagrangian algorithm for equality constrained optimization Optimization Methods and Software ISSN: 1055-6788 (Print) 1029-4937 (Online) Journal homepage: http://www.tandfonline.com/loi/goms20 A globally and quadratically convergent primal dual augmented Lagrangian

More information

Numerical Methods for Differential Equations Mathematical and Computational Tools

Numerical Methods for Differential Equations Mathematical and Computational Tools Numerical Methods for Differential Equations Mathematical and Computational Tools Gustaf Söderlind Numerical Analysis, Lund University Contents V4.16 Part 1. Vector norms, matrix norms and logarithmic

More information

Gradient Descent. Dr. Xiaowei Huang

Gradient Descent. Dr. Xiaowei Huang Gradient Descent Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Three machine learning algorithms: decision tree learning k-nn linear regression only optimization objectives are discussed,

More information

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Newton s Method. Ryan Tibshirani Convex Optimization /36-725 Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x

More information

A First Order Method for Finding Minimal Norm-Like Solutions of Convex Optimization Problems

A First Order Method for Finding Minimal Norm-Like Solutions of Convex Optimization Problems A First Order Method for Finding Minimal Norm-Like Solutions of Convex Optimization Problems Amir Beck and Shoham Sabach July 6, 2011 Abstract We consider a general class of convex optimization problems

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

Static unconstrained optimization

Static unconstrained optimization Static unconstrained optimization 2 In unconstrained optimization an objective function is minimized without any additional restriction on the decision variables, i.e. min f(x) x X ad (2.) with X ad R

More information

Control, Stabilization and Numerics for Partial Differential Equations

Control, Stabilization and Numerics for Partial Differential Equations Paris-Sud, Orsay, December 06 Control, Stabilization and Numerics for Partial Differential Equations Enrique Zuazua Universidad Autónoma 28049 Madrid, Spain enrique.zuazua@uam.es http://www.uam.es/enrique.zuazua

More information