Institute of Eng. and Comp. Mechanics Optimization of Mechanical Systems Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A1.1.

Size: px
Start display at page:

Download "Institute of Eng. and Comp. Mechanics Optimization of Mechanical Systems Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A1.1."

Transcription

1 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A1.1 Motivational Example 1 In medieval times catapults where used in the siege of castles. These catapults where capable of throwing heavy stones over a distance of several hundred meters. The catapult consists of an undercarriage, a boom with counter mass and the stone. The stone is connected to the boom by a rope. Optimize! u l s boom with counter weight stone h undercarriage p

2 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A1.2 Motivational Example 2 For different TVs the table below summarizes the screen diagonal and price. LCD TV Diagonal cm] Price Euro] A Sony KDL-26U B Sharp LC-26 D44 E-BK C Sharp LC-32 D44 EBK D Philips 32PFL9603D/ E Samsung LE-37R86BD F Panasonic TXLZD70F G Sony KDL-40V H Samsung LE40F86BD I Panasonic TXLZD70F J Panasonic TH-42PZ70E K Pioneer PDP-4280XA L Sharp LC-52 X 20 E M Samsung LE 52 A 759 R a.) Order the TVs with respect of the screen diameters. Find the TV with the largest screen diagonal. b.) Order the TVs with respect of the price. Find the TV with the minimal price. c.) Which TV is a compromise in respect of both criteria?

3 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A2 Example: Optimization of the rear suspension of a car technical system mechanical model M : car body mass m : wheel and axel mass c : stiffness of suspension d : damping of suspension c W : stiffness of tire s : street profile mathematical model Mx c(x x W ) d(x x w) mx w c(x x w ) + d(x x w) c w (x w s(t))

4 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A3 Notation a R, b R m, C R m n 1) derivative with respect to a scalar x R matrix notation index notation 2) derivative with respect to a vector x R n 3) derivative of a scalar product b x x R m b i const 4) derivative of a quadratic form x A x A A T R n n A ij const x R n

5 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A4.1 Graph for f(x 1, x 2 ) x 1 x 2 + ln (e x 2) Automatic Differentiation x 6 x 5 f x 6 x 5 + x 4 x 6 x 4 x 5 x 1 x 5 x 1 x 2 x 5 x 2 x 1 x 2 x 4 ln x 3 x 3 e x 2 x 3 x 2 x 4 x 3 Forward mode: x 1 ], x 2 ] x 3, x 3 x 3 x x j j j J 3 {2} x 3 x 2 x 2 ] x 4, x 4 x j j J 4 { } x 5, x 5 x j j J 5 { } ] x 6, x 6 x j j J 6 { } + ] + ]

6 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A4.2 Reverse mode: x 3, x 4, x 5, f x 6 x 6 x 6, x 6 x 5 x j x j x 5 j I 5 {6}, x 4 j I { }, x 3 j I { }, x 2 j I { } +, x 1 j I { }, f df dx x 1 x 1 ] ]. f x 2 x 2 ]

7 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A4.3

8 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A5.1 Automatic differentiation For optimization purposes the sensitivities of the function f(x) x 1 sin(x 2 ) 5x 3, with x x 1, x 2 x 3 ], must be calculated. a) Complete the graph for the function f(x) and write down the derivatives along the arcs. x 4 x 4 x 3 x 1 x 2 x 3

9 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A5.2 Firstly, use the forward mode to compute the gradient of the function. b) Give the index sets J i for your introduced intermediate and output quantities. c) Compute the gradients for all input and intermediate quantities. d) Give the gradient of the function f(x) in terms of the input quantities x x 1, x 2 x 3 ].

10 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A5.3 Use also the reverse mode to compute the gradient of the function. e) Give the index sets I i for the input quantities and the introduced intermediate quantities. f) Compute the scalar gradients x j for the output quantity and intermediate quantities. g) Give the gradient of the function f(x) in terms of the input quantities x x 1, x 2 x 3 ]. h) Check your result by direct differentiation of the function f(x).

11 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A6 Minima of the Rosenbrock-function A common testing example for optimization algorithms is the Rosenbrock-function f(p) 100(p 2 p 1 2 ) 2 + (1 p 1 ) 2 For this function many optimization algorithms show convergence problems. This is due to the slight curvature of the valley and the very slowly decreasing bottom of the valley. a) Complete the gradient and the Hessian matrix for an arbitrary point p. f(p) ] 2 f(p) ] b) Give the necessary condition of first order for minima. Compute from this a potential minimizer p. c) Does p satisfy the necessary condition of second order?

12 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A7.1 Minimization of a Quadratic Function Minimize the quadratic function f(p) 10p p 2 2. Analytical Solution Using the gradient and the Hessian matrix, f(p), 2 f(p), ] ] we get the solution from the necessary conditions for a local minimizer, i.e.,. Here, the solution p is also a global minimizer of the function. Newton Method Starting from the initial point p (0) 0.1, 1] the search direction is s (0) ( 2 f (0) ) 1 f (0) Along the corresponding line ] ] ]. p(α) p (0) + αs (0), the criterion function is ] f(α) f(p(α)). The minimum along this line follows from f! 0 α (0)

13 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A7.2 which yields the improved solution p (1) p(α (0) ) and so we got the minimizer in the first step. Conjugate Gradient Method Starting from the initial point p (0) 0.1, 1] we get the following iteration steps: ν p (v) f (v) exact line search 0 s (0) f (0) ] ] ] p(α) ] f(α) f! 0 α (0) 1 s (1) f (1) + f (1) 2 f (0) 2 s(0) ] ] ] p(α) ] f(α) 2 0 ] convergence in two steps 0 f! 0 α (1)

14 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A8.1 Minimization of a Quadratic Function Minimize the quadratic function f(p) 10p p 2 2. Quasi Newton Method Starting from the initial point p (0) 0.1, 1] we get the following iteration steps: ν p (v) f (v) exact line search 0 H (0) E ] ] ] s (0) H (0) f (0) ] p(α) ] f(α) f! 0 α (0) 1 ] ] δp p (1) p (0) ] γ f (1) f (0) ] δp γ, H (0) γ, γ H (0) γ

15 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A8.2 ν p (v) f (v) exact line search H (1) H (0) δpδp + H (1) ] s (1) H (1) f (1) ] p(α) ] f(α) f! 0 α (0) 2 0 ] convergence in two steps 0

16 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A9.1 Constrained Minimization of a Quadratic Function Minimize the quadratic function f(p) p p 2 2 subject to the constraint p 1 p 2 4 or written in the standard form h 1 (p) 0 Graphical Solution: For this simple problem, the solution can be found graphically in the design parameter space. The dimension of the design parameter space is. Plot the contour lines of the function f and the constraint h 1 and mark the optimal solution p.

17 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A9.2 Analytical Solution: The inequality constraint can be handled as an equality constraint by introducing the offset (slack) variable u 1 h 1 (p) 0 The Lagrange function for an optimization problem with constraints generally reads as It follows L(p, μ, u) f(p) μ_i(h i + u i 2 ) m i1 L(p, μ 1, u 1 ). The necessary condition of 1 st order for a minimum is L(p, μ 1, u 1 )! following four equations 0, which leads to the L p! ] ] L μ 1! L u 1! From the last equation we can distinguish two cases, either or.

18 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A9.3 Case 1: If μ 1 0, the constraint is not considered and therefore inactive. This is only true, if the constraint is fulfilled. It follows from L(p, μ 1, u 1 )! 0 p 1, p 2 and This last equation cannot be satisfied and subsequently the constraint is not fulfilled at the minimum of f at p ]and case 1 is not true. Case 2: If u 1 0, the distance from the limit of the constraint is 0 and at p we have an active constraint. The condition L(p, μ 1, u 1 )! 0 leaves us with three equations and it follows p 1, p 2, μ 1.

19 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A10.1 Constrained Minimization of a Nonlinear Function Minimize the nonlinear objective function f(p) (p p 2 ) 2 100p 1 subject to the equality constraint g(p) p p Solution by the Lagrange Newton Method: The Lagrange function for this optimization problem can be written as L(p, λ) λ. The Karush-Kuhn-Tucker (KKT) equations represent necessary conditions for an optimum. They can be written as L p! ] ] L λ! These nonlinear equations can be solved by the Newton-Raphson method. This yields the update equation of the Lagrange-Newton method 2 ( g(p) T p ) δp (v+1) δf(p) δp g(p) p ] p (v), λ (v) A(p,λ) λ (v+1) ] g(p) ] p (v) b(p)

20 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A10.2 With the objective and constraint function given above, the matrix A and the right hand side b follows as 30p A(p, λ) 200 2λ 2p 2, 0 ] b(p) 20p 1 3. ] Starting with p (0) 1 0 ] and λ (0) 0 the update equation reads as 30 δp (1) ] λ (1). ] 8 ]

21 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A10.3 From these three scalar equations, the updated optimization variables and the new Lagrange multiplier are given by p (1) p (0) + δp (1) ], λ (1) The next iteration yields p (2) λ (2) 2240, 2.6 ] ] ] which results in p (2) ] and λ (2) The following table summarizes the first 10 iterations of the Lagrange-Newton method: Iteration p 1 p 2 λ f g 0 1,0000E+00 0,0000E+00 0,0000E+00-9,9000E+01-8,0000E ,0000E+00-1,3000E+00-2,6000E+01 1,2044E+04 1,7690E ,7348E+00 6,3777E-01-3,1944E+02 3,0458E+03 5,3556E ,1496E+00-1,3357E-01 2,5548E+01 5,7946E+02 9,3751E ,8840E+00-2,8855E+00-1,1348E+02-2,6471E+02 7,6436E ,5959E+00-1,8490E+00-3,4899E+01-2,5859E+02 1,1574E ,5655E+00-1,5787E+00-1,0998E+01-2,5534E+02 7,3970E ,5667E+00-1,5533E+00-8,8899E+00-2,5478E+02 6,4691E ,5667E+00-1,5531E+00-8,8713E+00-2,5477E+02 4,6864E ,5667E+00-1,5531E+00-8,8713E+00-2,5477E+02 3,5527E ,5667E+00-1,5531E+00-8,8713E+00-2,5477E+02 1,0011E-15 Draw the points p (0) to p (4) and the final point p (10) in the contour plot on the next page. Another solution process was started with the initial point p (0) 1 following table summarizes the corresponding optimization history: 0] and λ (0) 0. The

22 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A10.4 Iteration p 1 p 2 λ f g 0-1,0000E+00 0,0000E+00 0,0000E+00 1,0100E+02-8,0000E ,0000E+00 1,3000E+00 7,4000E+01 1,3044E+04 1,7690E ,7693E+00-7,7022E-01-1,9290E+02 4,1289E+03 5,8005E ,2153E+00 2,8425E-01 8,7719E+01 1,2455E+03 1,4188E ,6628E+00 4,0384E+00-3,0435E+02 7,2870E+02 1,4399E ,3117E+00 2,4872E+00-8,3576E+01 3,8788E+02 2,5296E ,4700E+00 1,8314E+00-8,2745E+00 2,5753E+02 4,5508E ,4825E+00 1,6904E+00 8,1362E+00 2,5082E+02 2,0039E ,4838E+00 1,6826E+00 8,9300E+00 2,5064E+02 6,3143E ,4838E+00 1,6825E+00 8,9323E+00 2,5064E+02 4,5733E ,4838E+00 1,6825E+00 8,9323E+00 2,5064E+02-3,5527E-15 Draw the points p (0) to p (4) and the final point p (10) of this run in the contour plot, too. Contour plot of the constrained problem: As we can see, the solution depends on the starting point p (10). Visit the webpage and try to solve the problem with the stochastic particle swarm optimizer which is available there. What can you observe? What are the differences with respect to the Lagrange-Newton method?

23 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A11.1 Application: Optimization of a Double Pendulum An actively controlled double pendulum is described by the generalized coordinates y α β]. Without the torque actuator T, the dynamic behavior is described by the nonlinear equations of motion M(t, y) y + k(t, y, y ) q(t, y, y ) and the initial conditions are y(0) y 0, y (0) 0. Linearization of the equations of motion about the equilibrium position results in y(t) 0 + η(t), y (t) 0 + η (t), y (t) 0 + η (t) and the following system of linear equations where M(t) η (t) + P(t) η (t) + Q(t) η(t) h(t), M 2mL2 ml mgL 0 ml 2 2 ], P ], Q ml mgl ]. The generalized force vector q has to be considered with the control torque T(t) of the torque actuator h(t) T(t) 0 ] and the PD control law is T(t) k 1 α k 2 α

24 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A11.2 By an optimization process the control parameters k 1 and k 2 are to be optimized in such a way, that the vibrations of the pendulum, evaluated by the integral of y y within the first 10 s, are optimally damped. In order to ensure the stability of the system, both of the control parameters must not be negative. First we formulate the problem in standard form: design variables p ʼ ] kinetics initial conditions φ 0 0, φ 0 0. ] ] final condition H 1 0 criteria functions Ψ 1 Ψ 2 Ψ 3 optimization criterion inequality constraints f h 1 The constrained optimization task then reads as h 2 find, where and

25 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A11.3 The described optimization task is solved by a SQP method. The parameters of the double pendulum are m 1kg, L 0.3m, y ], and the initial design variable vector is p(0) ]. The optimization algorithm needs 10 gradient evaluations and 16 function evaluations in order to find the optimizer p ]. The optimization history shows the convergence after 10 iterations: The optimal dynamic behaviour shows a significantly improved damping in comparison to the passive and the initial design:

26 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A12.1 Multicriteria Optimization of a Truss For the presented truss the cross sections A 1 and A 2 of the two bars are to be designed. Simultaneously, the overall mass and the vertical displacement of the point P shall be minimized, while an external load F is applied at P. This is a typical problem of multicriteria optimization, since the overall mass corresponds to the costs of material and the vertical displacement is a measure for the function of the framework. Using the density ρ the overall mass is M For the computation of the vertical displacement the mass of the truss can be neglected in comparison to the external load. Using finite element modelling we get for the displacement y of the joint K y q where K E 2 2l A 2 A 2 ], y u A 2 2 2A 1 + A 2 v ], q 0 F ] which yields the vertical displacement v If A is a given minimal cross section of the bars, we get dimensionless design variables by normalization: p 1, p 2 - and the dimensionless optimization criteria f 1 M ρla, f 2 v Fl AE

27 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A12.2 If the possible cross section has a maximum of αa, then we get the vector optimization problem opt p P ] where P {p R 2 } Sketch the problem in the design space and in the criteria space for α 4. f(p)

28 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A12.3 From these sketches we get the Edgeworth Pareto optimal solutions P P { ], } For the numerical determination of the EP optimal solutions we generally formulate several scalar optimization problems. Formulate the following scalar optimization problems and determine the corresponding solution graphically. 1) Weighted Criteria Method (Weighting factors:w 1 w 2 1 2) min p P 2) Distance Method (Minimization of the Euklidean distance to the ideal solution f 0 ʼ ]) min p P

29 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A12.4 3) Hierarchical Method If the minimal vertical displacement is more important than the minimal overall mass of the truss, then the priorities are l 1 and l 2. In step 1, the minimization of the vertical displacement only leads to f 2 min p P 0 f 2(p) where P 0 In step 2, we try to additionally minimize the overall mass. For this, we have to allow some increase of the vertical displacement, e.g. ε 1 100%. The second optimization step then reads f 1 min p P 1 f 1(p) where P 1 { }

30 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A13.1 MATLAB basics Numerical Tools Optimization with MATLAB The program MATLAB is a commercially available program. It uses a numeric representation of vectors and matrices and allows a simple programming of matrix operations and provides many different methods to analyse systems. Many so-called Toolboxes are available to extend the functionality of MATLAB. The commands can be entered directly at the MATLAB prompt and executed immediately. Alternatively the commands can be summarized in a text file <name>.m. This socalled m-file is simply loaded and executed in MATLAB by typing <name> at the MATLAB prompt. helpdesk a ] b ]' A 1 2 4; ; ] starts online help defines a row vector defines a column vector defines matrix A transpose of matrix A A(2,3) A(:,3) A(1,:) zeros(n,m) ones(n,m) diag(1 2 3]) c A * b c A \ b inv(a) matrix element A23 3 rd column of matrix A 1 st row of matrix A defines a zero matrix defines a matrix of ones defines a diagonal matrix matrix-vector multiplication solves linear system inverse of a matrix plot(t,x, g ),hold on; EV, ew ] eig(a) matrix EV of eigenvectors and vector ew of eigenvalues of matrix A plot vector x over vector t For the numerical integration MATLAB provides two explicit single-step methods (ode23, ode45), one implicit single-step method (ode23s), an Adams-Bashfort- Moulton predictor corrector method (ode113) and an implicit numerical differentiation formula (ode15s). All methods have implemented a step size control. t,x] ode23('<name>',tspan,x0) simulation of a system of ordinary differential equations, given in the form dx/dt f(t,x). Outputs are the vector t of time steps and

31 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A13.2 the matrix X of state trajectories. The number of columns in X corresponds to the number of states and the number of rows to the number of time steps. Input is the <name> of a function specifying dx/dt, a (2x1) vector tspan t0, tend] contains the initial and the final time and the initial state vector is x0 1 ( t ) x ( t ) x ( t ) x ( t ) ]. x The function <name> must be written in a separate file <name.m> with the corresponding file name and must follow the syntax: function dx <name>(t, x)... dx... Example: Linearized double pendulum function dx doublependulum(t,x) M1 1.0; IZZ 0.0; GE 9.81; L 0.3; M 2.0*M1*L^2+IZZ M1*L^2; M1*L^2 M1*L^2+IZZ ]; Q 2*GE*M1*L 0; 0 GE*M1*L ]; A zeros(2) diag( 1 1]); -inv(m)*q zeros(2)]; dx A*x; MATLAB Optimization Toolbox The MATALAB Optimization Toolbox provides deterministic optimization algorithms for linear and non-nonlinear problems. The toolbox includes different solution methods for minimization problems, equations of nonlinear systems (fsolve, fzero) and least squares problems (curve fitting). Some important commands (described for medium scale optimization problems): fminsearch fminunc Finds minimum of unconstrained multivariable function using the derivative-free Simplex Search Method. Finds minimum of unconstrained multivariable function. The algorithm uses a Quasi-Newton method with a cubic line search procedure. This Quasi-Newton method uses the Broyden-Fletcher-Goldfarb-Shanno formula for updating the approximation of the Hessian matrix or the Davidson- Fletcher-Powell formula for approximating the inverse Hessian matrix.

32 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A13.3 fmincon Finds minimum of constrained nonlinear multivariable function. The algorithm uses a SQP method. During each iteration step an estimate of the Hessian of the Lagrangian is updated and a line search is performed. Example: Multicriteria Optimization of a truss structure using weighted criteria (A11) Main m-file truss.m for optimization: clear all; global w1 w2 % bounds of design variables alpha 3; A 1 0; 0 1; -1 0 ; 0-1]; b alpha alpha -1-1]'; % start values x0 1 1]; % options for optimization algorithm options optimset('largescale','off'); % weighting factors w1 0.3; w2 1-w1; % optimization x,fval] fmincon(@criteria,x0,a,b,],],],],],options) M-file criteria.m for function evaluation: function f criteria(x) global w1 w2 % calculation of optimization criteria % criteria f1 and f2 f1 x(1)+sqrt(2)*x(2); f2 1/x(1); % scalar optimization criteria f w1*f1+w2*f2; Example: Computation of the Pareto-Front of a truss structure (A11) The file truss.m is extended in order to compute the entire Pareto-Front. Thereby the weighted criteria method with variable weights is used. clear all; global w1 w2 alpha 3; % bounds of design variables A 1 0; 0 1; -1 0 ; 0-1]; b alpha alpha -1-1]'; % start values x0 1 1]; % options for optimization algorithm

33 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A13.4 options optimset('largescale','off'); for i1:51; % weighting factors w1 (i-1)/50; w2 1-w1; % optimization x(:,i),fval] fmincon(@criteria,x0,a,b,... ],],],],],options); % use previous results as next starting values x0 x(:,i); % criteria f1 and f2 f1(i) x(1,i)+sqrt(2)*x(2,i); f2(i) 1/x(1,i); end figure(1) subplot(1,2,1) plot(1 alpha],1 1],'b'), hold on; plot(1 1],1 alpha],'b'); plot(1 alpha],alpha alpha],'b'); plot(alpha alpha],1 alpha],'b'); plot(x(1,:),x(2,:),'r+'); axis(0.8 alpha alpha+0.2]) xlabel('p_1') ylabel('p_2') subplot(1,2,2) plot(f1,f2,'r+'), hold on; axis( ]) xlabel('f_1') ylabel('f_2') Figure 1: Pareto-Front of the truss structure. What problem do you see? MATLAB Genetic Algorithm and Direct Search Toolbox The MATLAB Genetic Algorithm and Direct Search Toolbox is a collection of stochastic optimization algorithms. This toolbox includes routines for solving optimization problems using Direct Search, Genetic Algorithm and Simulated Annealing.

34 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A13.5 Optimization with EXCEL For optimizing nonlinear problems EXCEL uses the Generalized Reduced Gradient (GRG2) algorithm developed by Leon Lasdon and Allan Waren. Other solvers can be installed, see e.g.

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

Optimization Methods

Optimization Methods Optimization Methods Decision making Examples: determining which ingredients and in what quantities to add to a mixture being made so that it will meet specifications on its composition allocating available

More information

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Multidimensional Unconstrained Optimization Suppose we have a function f() of more than one

More information

4M020 Design tools. Algorithms for numerical optimization. L.F.P. Etman. Department of Mechanical Engineering Eindhoven University of Technology

4M020 Design tools. Algorithms for numerical optimization. L.F.P. Etman. Department of Mechanical Engineering Eindhoven University of Technology 4M020 Design tools Algorithms for numerical optimization L.F.P. Etman Department of Mechanical Engineering Eindhoven University of Technology Wednesday September 3, 2008 1 / 32 Outline 1 Problem formulation:

More information

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α

More information

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality

More information

Lecture V. Numerical Optimization

Lecture V. Numerical Optimization Lecture V Numerical Optimization Gianluca Violante New York University Quantitative Macroeconomics G. Violante, Numerical Optimization p. 1 /19 Isomorphism I We describe minimization problems: to maximize

More information

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints Instructor: Prof. Kevin Ross Scribe: Nitish John October 18, 2011 1 The Basic Goal The main idea is to transform a given constrained

More information

Optimization Methods

Optimization Methods Optimization Methods Categorization of Optimization Problems Continuous Optimization Discrete Optimization Combinatorial Optimization Variational Optimization Common Optimization Concepts in Computer Vision

More information

Scientific Computing: Optimization

Scientific Computing: Optimization Scientific Computing: Optimization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 March 8th, 2011 A. Donev (Courant Institute) Lecture

More information

Numerical Optimization: Basic Concepts and Algorithms

Numerical Optimization: Basic Concepts and Algorithms May 27th 2015 Numerical Optimization: Basic Concepts and Algorithms R. Duvigneau R. Duvigneau - Numerical Optimization: Basic Concepts and Algorithms 1 Outline Some basic concepts in optimization Some

More information

Algorithms for Constrained Optimization

Algorithms for Constrained Optimization 1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic

More information

Review for Exam 2 Ben Wang and Mark Styczynski

Review for Exam 2 Ben Wang and Mark Styczynski Review for Exam Ben Wang and Mark Styczynski This is a rough approximation of what we went over in the review session. This is actually more detailed in portions than what we went over. Also, please note

More information

2.3 Linear Programming

2.3 Linear Programming 2.3 Linear Programming Linear Programming (LP) is the term used to define a wide range of optimization problems in which the objective function is linear in the unknown variables and the constraints are

More information

Optimization and Root Finding. Kurt Hornik

Optimization and Root Finding. Kurt Hornik Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding

More information

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09 Numerical Optimization 1 Working Horse in Computer Vision Variational Methods Shape Analysis Machine Learning Markov Random Fields Geometry Common denominator: optimization problems 2 Overview of Methods

More information

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition) NONLINEAR PROGRAMMING (Hillier & Lieberman Introduction to Operations Research, 8 th edition) Nonlinear Programming g Linear programming has a fundamental role in OR. In linear programming all its functions

More information

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be

More information

ECS550NFB Introduction to Numerical Methods using Matlab Day 2

ECS550NFB Introduction to Numerical Methods using Matlab Day 2 ECS550NFB Introduction to Numerical Methods using Matlab Day 2 Lukas Laffers lukas.laffers@umb.sk Department of Mathematics, University of Matej Bel June 9, 2015 Today Root-finding: find x that solves

More information

Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

More information

Numerical Methods for the Solution of Differential Equations

Numerical Methods for the Solution of Differential Equations Numerical Methods for the Solution of Differential Equations Markus Grasmair Vienna, winter term 2011 2012 Analytical Solutions of Ordinary Differential Equations 1. Find the general solution of the differential

More information

Interior-Point Methods for Linear Optimization

Interior-Point Methods for Linear Optimization Interior-Point Methods for Linear Optimization Robert M. Freund and Jorge Vera March, 204 c 204 Robert M. Freund and Jorge Vera. All rights reserved. Linear Optimization with a Logarithmic Barrier Function

More information

Optimization of Mechanical Systems

Optimization of Mechanical Systems Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 M0 Optimization of Mechanical Systems Prof. Dr.-Ing. Prof. E.h. P. Eberhard (peter.eberhard@itm.uni-stuttgart.de) M.Sc. Elizaveta Shishova (elizaveta.shishova@itm.uni-stuttgart.de)

More information

IE 5531: Engineering Optimization I

IE 5531: Engineering Optimization I IE 5531: Engineering Optimization I Lecture 12: Nonlinear optimization, continued Prof. John Gunnar Carlsson October 20, 2010 Prof. John Gunnar Carlsson IE 5531: Engineering Optimization I October 20,

More information

SMO Algorithms for Support Vector Machines without Bias Term

SMO Algorithms for Support Vector Machines without Bias Term Institute of Automatic Control Laboratory for Control Systems and Process Automation Prof. Dr.-Ing. Dr. h. c. Rolf Isermann SMO Algorithms for Support Vector Machines without Bias Term Michael Vogt, 18-Jul-2002

More information

Multivariate Newton Minimanization

Multivariate Newton Minimanization Multivariate Newton Minimanization Optymalizacja syntezy biosurfaktantu Rhamnolipid Rhamnolipids are naturally occuring glycolipid produced commercially by the Pseudomonas aeruginosa species of bacteria.

More information

Numerical Optimization. Review: Unconstrained Optimization

Numerical Optimization. Review: Unconstrained Optimization Numerical Optimization Finding the best feasible solution Edward P. Gatzke Department of Chemical Engineering University of South Carolina Ed Gatzke (USC CHE ) Numerical Optimization ECHE 589, Spring 2011

More information

Nonlinear Programming (Hillier, Lieberman Chapter 13) CHEM-E7155 Production Planning and Control

Nonlinear Programming (Hillier, Lieberman Chapter 13) CHEM-E7155 Production Planning and Control Nonlinear Programming (Hillier, Lieberman Chapter 13) CHEM-E7155 Production Planning and Control 19/4/2012 Lecture content Problem formulation and sample examples (ch 13.1) Theoretical background Graphical

More information

Multidisciplinary System Design Optimization (MSDO)

Multidisciplinary System Design Optimization (MSDO) Multidisciplinary System Design Optimization (MSDO) Numerical Optimization II Lecture 8 Karen Willcox 1 Massachusetts Institute of Technology - Prof. de Weck and Prof. Willcox Today s Topics Sequential

More information

Numerical Methods I Solving Nonlinear Equations

Numerical Methods I Solving Nonlinear Equations Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)

More information

Constrained optimization: direct methods (cont.)

Constrained optimization: direct methods (cont.) Constrained optimization: direct methods (cont.) Jussi Hakanen Post-doctoral researcher jussi.hakanen@jyu.fi Direct methods Also known as methods of feasible directions Idea in a point x h, generate a

More information

The roots are found with the following two statements. We have denoted the polynomial as p1, and the roots as roots_ p1.

The roots are found with the following two statements. We have denoted the polynomial as p1, and the roots as roots_ p1. Part II Lesson 10 Numerical Analysis Finding roots of a polynomial In MATLAB, a polynomial is expressed as a row vector of the form [an an 1 a2 a1 a0]. The elements ai of this vector are the coefficients

More information

Real-time Constrained Nonlinear Optimization for Maximum Power Take-off of a Wave Energy Converter

Real-time Constrained Nonlinear Optimization for Maximum Power Take-off of a Wave Energy Converter Real-time Constrained Nonlinear Optimization for Maximum Power Take-off of a Wave Energy Converter Thomas Bewley 23 May 2014 Southern California Optimization Day Summary 1 Introduction 2 Nonlinear Model

More information

Unconstrained Multivariate Optimization

Unconstrained Multivariate Optimization Unconstrained Multivariate Optimization Multivariate optimization means optimization of a scalar function of a several variables: and has the general form: y = () min ( ) where () is a nonlinear scalar-valued

More information

Written Examination

Written Examination Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes

More information

NonlinearOptimization

NonlinearOptimization 1/35 NonlinearOptimization Pavel Kordík Department of Computer Systems Faculty of Information Technology Czech Technical University in Prague Jiří Kašpar, Pavel Tvrdík, 2011 Unconstrained nonlinear optimization,

More information

Quasi-Newton Methods

Quasi-Newton Methods Newton s Method Pros and Cons Quasi-Newton Methods MA 348 Kurt Bryan Newton s method has some very nice properties: It s extremely fast, at least once it gets near the minimum, and with the simple modifications

More information

Numerical Optimization of Partial Differential Equations

Numerical Optimization of Partial Differential Equations Numerical Optimization of Partial Differential Equations Part I: basic optimization concepts in R n Bartosz Protas Department of Mathematics & Statistics McMaster University, Hamilton, Ontario, Canada

More information

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL) Part 4: Active-set methods for linearly constrained optimization Nick Gould RAL fx subject to Ax b Part C course on continuoue optimization LINEARLY CONSTRAINED MINIMIZATION fx subject to Ax { } b where

More information

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem: CDS270 Maryam Fazel Lecture 2 Topics from Optimization and Duality Motivation network utility maximization (NUM) problem: consider a network with S sources (users), each sending one flow at rate x s, through

More information

CHAPTER 2: QUADRATIC PROGRAMMING

CHAPTER 2: QUADRATIC PROGRAMMING CHAPTER 2: QUADRATIC PROGRAMMING Overview Quadratic programming (QP) problems are characterized by objective functions that are quadratic in the design variables, and linear constraints. In this sense,

More information

Static unconstrained optimization

Static unconstrained optimization Static unconstrained optimization 2 In unconstrained optimization an objective function is minimized without any additional restriction on the decision variables, i.e. min f(x) x X ad (2.) with X ad R

More information

Appendix A Taylor Approximations and Definite Matrices

Appendix A Taylor Approximations and Definite Matrices Appendix A Taylor Approximations and Definite Matrices Taylor approximations provide an easy way to approximate a function as a polynomial, using the derivatives of the function. We know, from elementary

More information

Nonlinear Optimization for Optimal Control

Nonlinear Optimization for Optimal Control Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]

More information

A Trust-region-based Sequential Quadratic Programming Algorithm

A Trust-region-based Sequential Quadratic Programming Algorithm Downloaded from orbit.dtu.dk on: Oct 19, 2018 A Trust-region-based Sequential Quadratic Programming Algorithm Henriksen, Lars Christian; Poulsen, Niels Kjølstad Publication date: 2010 Document Version

More information

Lecture 18: Optimization Programming

Lecture 18: Optimization Programming Fall, 2016 Outline Unconstrained Optimization 1 Unconstrained Optimization 2 Equality-constrained Optimization Inequality-constrained Optimization Mixture-constrained Optimization 3 Quadratic Programming

More information

What s New in Active-Set Methods for Nonlinear Optimization?

What s New in Active-Set Methods for Nonlinear Optimization? What s New in Active-Set Methods for Nonlinear Optimization? Philip E. Gill Advances in Numerical Computation, Manchester University, July 5, 2011 A Workshop in Honor of Sven Hammarling UCSD Center for

More information

Gradient Descent. Dr. Xiaowei Huang

Gradient Descent. Dr. Xiaowei Huang Gradient Descent Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Three machine learning algorithms: decision tree learning k-nn linear regression only optimization objectives are discussed,

More information

Neural Network Training

Neural Network Training Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification

More information

Numerical optimization

Numerical optimization Numerical optimization Lecture 4 Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Numerical geometry of non-rigid shapes Stanford University, Winter 2009 2 Longest Slowest Shortest Minimal Maximal

More information

MATH2070 Optimisation

MATH2070 Optimisation MATH2070 Optimisation Nonlinear optimisation with constraints Semester 2, 2012 Lecturer: I.W. Guo Lecture slides courtesy of J.R. Wishart Review The full nonlinear optimisation problem with equality constraints

More information

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A

More information

Math 409/509 (Spring 2011)

Math 409/509 (Spring 2011) Math 409/509 (Spring 2011) Instructor: Emre Mengi Study Guide for Homework 2 This homework concerns the root-finding problem and line-search algorithms for unconstrained optimization. Please don t hesitate

More information

Math Ordinary Differential Equations

Math Ordinary Differential Equations Math 411 - Ordinary Differential Equations Review Notes - 1 1 - Basic Theory A first order ordinary differential equation has the form x = f(t, x) (11) Here x = dx/dt Given an initial data x(t 0 ) = x

More information

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems AMSC 607 / CMSC 764 Advanced Numerical Optimization Fall 2008 UNIT 3: Constrained Optimization PART 4: Introduction to Interior Point Methods Dianne P. O Leary c 2008 Interior Point Methods We ll discuss

More information

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Newton s Method. Ryan Tibshirani Convex Optimization /36-725 Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x

More information

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Robert M. Freund March, 2004 2004 Massachusetts Institute of Technology. The Problem The logarithmic barrier approach

More information

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23 Optimization: Nonlinear Optimization without Constraints Nonlinear Optimization without Constraints 1 / 23 Nonlinear optimization without constraints Unconstrained minimization min x f(x) where f(x) is

More information

Numerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems

Numerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems 1 Numerical optimization Alexander & Michael Bronstein, 2006-2009 Michael Bronstein, 2010 tosca.cs.technion.ac.il/book Numerical optimization 048921 Advanced topics in vision Processing and Analysis of

More information

In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written

In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written 11.8 Inequality Constraints 341 Because by assumption x is a regular point and L x is positive definite on M, it follows that this matrix is nonsingular (see Exercise 11). Thus, by the Implicit Function

More information

Optimization Methods for Circuit Design

Optimization Methods for Circuit Design Technische Universität München Department of Electrical Engineering and Information Technology Institute for Electronic Design Automation Optimization Methods for Circuit Design Compendium H. Graeb Version

More information

Numerical Optimization Algorithms

Numerical Optimization Algorithms Numerical Optimization Algorithms 1. Overview. Calculus of Variations 3. Linearized Supersonic Flow 4. Steepest Descent 5. Smoothed Steepest Descent Overview 1 Two Main Categories of Optimization Algorithms

More information

Outline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St

Outline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic

More information

CEE Computer Applications. Mathematical Programming (LP) and Excel Solver

CEE Computer Applications. Mathematical Programming (LP) and Excel Solver CEE 3804 - Computer Applications Mathematical Programming (LP) and Excel Solver Dr. Antonio A. Trani Professor of Civil and Environmental Engineering Virginia Polytechnic Institute and State University

More information

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren SF2822 Applied Nonlinear Optimization Lecture 9: Sequential quadratic programming Anders Forsgren SF2822 Applied Nonlinear Optimization, KTH / 24 Lecture 9, 207/208 Preparatory question. Try to solve theory

More information

MATH 4211/6211 Optimization Quasi-Newton Method

MATH 4211/6211 Optimization Quasi-Newton Method MATH 4211/6211 Optimization Quasi-Newton Method Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Quasi-Newton Method Motivation:

More information

Constrained Optimization

Constrained Optimization 1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange

More information

Lecture 3. Optimization Problems and Iterative Algorithms

Lecture 3. Optimization Problems and Iterative Algorithms Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex

More information

5 Quasi-Newton Methods

5 Quasi-Newton Methods Unconstrained Convex Optimization 26 5 Quasi-Newton Methods If the Hessian is unavailable... Notation: H = Hessian matrix. B is the approximation of H. C is the approximation of H 1. Problem: Solve min

More information

CHAPTER 12 TIME DOMAIN: MODAL STATE SPACE FORM

CHAPTER 12 TIME DOMAIN: MODAL STATE SPACE FORM CHAPTER 1 TIME DOMAIN: MODAL STATE SPACE FORM 1.1 Introduction In Chapter 7 we derived the equations of motion in modal form for the system in Figure 1.1. In this chapter we will convert the modal form

More information

Optimization Concepts and Applications in Engineering

Optimization Concepts and Applications in Engineering Optimization Concepts and Applications in Engineering Ashok D. Belegundu, Ph.D. Department of Mechanical Engineering The Pennsylvania State University University Park, Pennsylvania Tirupathi R. Chandrupatia,

More information

MATH 3795 Lecture 13. Numerical Solution of Nonlinear Equations in R N.

MATH 3795 Lecture 13. Numerical Solution of Nonlinear Equations in R N. MATH 3795 Lecture 13. Numerical Solution of Nonlinear Equations in R N. Dmitriy Leykekhman Fall 2008 Goals Learn about different methods for the solution of F (x) = 0, their advantages and disadvantages.

More information

MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year

MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2 1 Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year 2013-14 OUTLINE OF WEEK 8 topics: quadratic optimisation, least squares,

More information

MS&E 318 (CME 338) Large-Scale Numerical Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization 1 Origins Instructor: Michael Saunders Spring 2015 Notes 9: Augmented Lagrangian Methods

More information

STATE VARIABLE (SV) SYSTEMS

STATE VARIABLE (SV) SYSTEMS Copyright F.L. Lewis 999 All rights reserved Updated:Tuesday, August 05, 008 STATE VARIABLE (SV) SYSTEMS A natural description for dynamical systems is the nonlinear state-space or state variable (SV)

More information

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study International Journal of Mathematics And Its Applications Vol.2 No.4 (2014), pp.47-56. ISSN: 2347-1557(online) Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms:

More information

Algorithms for nonlinear programming problems II

Algorithms for nonlinear programming problems II Algorithms for nonlinear programming problems II Martin Branda Charles University Faculty of Mathematics and Physics Department of Probability and Mathematical Statistics Computational Aspects of Optimization

More information

Contents. Preface. 1 Introduction Optimization view on mathematical models NLP models, black-box versus explicit expression 3

Contents. Preface. 1 Introduction Optimization view on mathematical models NLP models, black-box versus explicit expression 3 Contents Preface ix 1 Introduction 1 1.1 Optimization view on mathematical models 1 1.2 NLP models, black-box versus explicit expression 3 2 Mathematical modeling, cases 7 2.1 Introduction 7 2.2 Enclosing

More information

Algorithms for constrained local optimization

Algorithms for constrained local optimization Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained

More information

MODERN CONTROL DESIGN

MODERN CONTROL DESIGN CHAPTER 8 MODERN CONTROL DESIGN The classical design techniques of Chapters 6 and 7 are based on the root-locus and frequency response that utilize only the plant output for feedback with a dynamic controller

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week

More information

Computational Finance

Computational Finance Department of Mathematics at University of California, San Diego Computational Finance Optimization Techniques [Lecture 2] Michael Holst January 9, 2017 Contents 1 Optimization Techniques 3 1.1 Examples

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information

Optimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29

Optimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29 Optimization Yuh-Jye Lee Data Science and Machine Intelligence Lab National Chiao Tung University March 21, 2017 1 / 29 You Have Learned (Unconstrained) Optimization in Your High School Let f (x) = ax

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428

More information

Ordinary differential equations - Initial value problems

Ordinary differential equations - Initial value problems Education has produced a vast population able to read but unable to distinguish what is worth reading. G.M. TREVELYAN Chapter 6 Ordinary differential equations - Initial value problems In this chapter

More information

Mathematical Economics. Lecture Notes (in extracts)

Mathematical Economics. Lecture Notes (in extracts) Prof. Dr. Frank Werner Faculty of Mathematics Institute of Mathematical Optimization (IMO) http://math.uni-magdeburg.de/ werner/math-ec-new.html Mathematical Economics Lecture Notes (in extracts) Winter

More information

FALL 2018 MATH 4211/6211 Optimization Homework 4

FALL 2018 MATH 4211/6211 Optimization Homework 4 FALL 2018 MATH 4211/6211 Optimization Homework 4 This homework assignment is open to textbook, reference books, slides, and online resources, excluding any direct solution to the problem (such as solution

More information

Applied Computational Economics Workshop. Part 3: Nonlinear Equations

Applied Computational Economics Workshop. Part 3: Nonlinear Equations Applied Computational Economics Workshop Part 3: Nonlinear Equations 1 Overview Introduction Function iteration Newton s method Quasi-Newton methods Practical example Practical issues 2 Introduction Nonlinear

More information

Chapter 2. Optimization. Gradients, convexity, and ALS

Chapter 2. Optimization. Gradients, convexity, and ALS Chapter 2 Optimization Gradients, convexity, and ALS Contents Background Gradient descent Stochastic gradient descent Newton s method Alternating least squares KKT conditions 2 Motivation We can solve

More information

Optimization Problems with Constraints - introduction to theory, numerical Methods and applications

Optimization Problems with Constraints - introduction to theory, numerical Methods and applications Optimization Problems with Constraints - introduction to theory, numerical Methods and applications Dr. Abebe Geletu Ilmenau University of Technology Department of Simulation and Optimal Processes (SOP)

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table

More information

Optimisation in Higher Dimensions

Optimisation in Higher Dimensions CHAPTER 6 Optimisation in Higher Dimensions Beyond optimisation in 1D, we will study two directions. First, the equivalent in nth dimension, x R n such that f(x ) f(x) for all x R n. Second, constrained

More information

minimize x subject to (x 2)(x 4) u,

minimize x subject to (x 2)(x 4) u, Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for

More information

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method. Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization

More information

WHY DUALITY? Gradient descent Newton s method Quasi-newton Conjugate gradients. No constraints. Non-differentiable ???? Constrained problems? ????

WHY DUALITY? Gradient descent Newton s method Quasi-newton Conjugate gradients. No constraints. Non-differentiable ???? Constrained problems? ???? DUALITY WHY DUALITY? No constraints f(x) Non-differentiable f(x) Gradient descent Newton s method Quasi-newton Conjugate gradients etc???? Constrained problems? f(x) subject to g(x) apple 0???? h(x) =0

More information

Numerical Methods for Embedded Optimization and Optimal Control. Exercises

Numerical Methods for Embedded Optimization and Optimal Control. Exercises Summer Course Numerical Methods for Embedded Optimization and Optimal Control Exercises Moritz Diehl, Daniel Axehill and Lars Eriksson June 2011 Introduction This collection of exercises is intended to

More information

Review of Optimization Basics

Review of Optimization Basics Review of Optimization Basics. Introduction Electricity markets throughout the US are said to have a two-settlement structure. The reason for this is that the structure includes two different markets:

More information

Survey of NLP Algorithms. L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA

Survey of NLP Algorithms. L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA Survey of NLP Algorithms L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA NLP Algorithms - Outline Problem and Goals KKT Conditions and Variable Classification Handling

More information

CE 191: Civil & Environmental Engineering Systems Analysis. LEC 17 : Final Review

CE 191: Civil & Environmental Engineering Systems Analysis. LEC 17 : Final Review CE 191: Civil & Environmental Engineering Systems Analysis LEC 17 : Final Review Professor Scott Moura Civil & Environmental Engineering University of California, Berkeley Fall 2014 Prof. Moura UC Berkeley

More information