Institute of Eng. and Comp. Mechanics Optimization of Mechanical Systems Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A1.1.
|
|
- Darcy May
- 5 years ago
- Views:
Transcription
1 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A1.1 Motivational Example 1 In medieval times catapults where used in the siege of castles. These catapults where capable of throwing heavy stones over a distance of several hundred meters. The catapult consists of an undercarriage, a boom with counter mass and the stone. The stone is connected to the boom by a rope. Optimize! u l s boom with counter weight stone h undercarriage p
2 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A1.2 Motivational Example 2 For different TVs the table below summarizes the screen diagonal and price. LCD TV Diagonal cm] Price Euro] A Sony KDL-26U B Sharp LC-26 D44 E-BK C Sharp LC-32 D44 EBK D Philips 32PFL9603D/ E Samsung LE-37R86BD F Panasonic TXLZD70F G Sony KDL-40V H Samsung LE40F86BD I Panasonic TXLZD70F J Panasonic TH-42PZ70E K Pioneer PDP-4280XA L Sharp LC-52 X 20 E M Samsung LE 52 A 759 R a.) Order the TVs with respect of the screen diameters. Find the TV with the largest screen diagonal. b.) Order the TVs with respect of the price. Find the TV with the minimal price. c.) Which TV is a compromise in respect of both criteria?
3 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A2 Example: Optimization of the rear suspension of a car technical system mechanical model M : car body mass m : wheel and axel mass c : stiffness of suspension d : damping of suspension c W : stiffness of tire s : street profile mathematical model Mx c(x x W ) d(x x w) mx w c(x x w ) + d(x x w) c w (x w s(t))
4 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A3 Notation a R, b R m, C R m n 1) derivative with respect to a scalar x R matrix notation index notation 2) derivative with respect to a vector x R n 3) derivative of a scalar product b x x R m b i const 4) derivative of a quadratic form x A x A A T R n n A ij const x R n
5 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A4.1 Graph for f(x 1, x 2 ) x 1 x 2 + ln (e x 2) Automatic Differentiation x 6 x 5 f x 6 x 5 + x 4 x 6 x 4 x 5 x 1 x 5 x 1 x 2 x 5 x 2 x 1 x 2 x 4 ln x 3 x 3 e x 2 x 3 x 2 x 4 x 3 Forward mode: x 1 ], x 2 ] x 3, x 3 x 3 x x j j j J 3 {2} x 3 x 2 x 2 ] x 4, x 4 x j j J 4 { } x 5, x 5 x j j J 5 { } ] x 6, x 6 x j j J 6 { } + ] + ]
6 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A4.2 Reverse mode: x 3, x 4, x 5, f x 6 x 6 x 6, x 6 x 5 x j x j x 5 j I 5 {6}, x 4 j I { }, x 3 j I { }, x 2 j I { } +, x 1 j I { }, f df dx x 1 x 1 ] ]. f x 2 x 2 ]
7 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A4.3
8 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A5.1 Automatic differentiation For optimization purposes the sensitivities of the function f(x) x 1 sin(x 2 ) 5x 3, with x x 1, x 2 x 3 ], must be calculated. a) Complete the graph for the function f(x) and write down the derivatives along the arcs. x 4 x 4 x 3 x 1 x 2 x 3
9 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A5.2 Firstly, use the forward mode to compute the gradient of the function. b) Give the index sets J i for your introduced intermediate and output quantities. c) Compute the gradients for all input and intermediate quantities. d) Give the gradient of the function f(x) in terms of the input quantities x x 1, x 2 x 3 ].
10 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A5.3 Use also the reverse mode to compute the gradient of the function. e) Give the index sets I i for the input quantities and the introduced intermediate quantities. f) Compute the scalar gradients x j for the output quantity and intermediate quantities. g) Give the gradient of the function f(x) in terms of the input quantities x x 1, x 2 x 3 ]. h) Check your result by direct differentiation of the function f(x).
11 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A6 Minima of the Rosenbrock-function A common testing example for optimization algorithms is the Rosenbrock-function f(p) 100(p 2 p 1 2 ) 2 + (1 p 1 ) 2 For this function many optimization algorithms show convergence problems. This is due to the slight curvature of the valley and the very slowly decreasing bottom of the valley. a) Complete the gradient and the Hessian matrix for an arbitrary point p. f(p) ] 2 f(p) ] b) Give the necessary condition of first order for minima. Compute from this a potential minimizer p. c) Does p satisfy the necessary condition of second order?
12 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A7.1 Minimization of a Quadratic Function Minimize the quadratic function f(p) 10p p 2 2. Analytical Solution Using the gradient and the Hessian matrix, f(p), 2 f(p), ] ] we get the solution from the necessary conditions for a local minimizer, i.e.,. Here, the solution p is also a global minimizer of the function. Newton Method Starting from the initial point p (0) 0.1, 1] the search direction is s (0) ( 2 f (0) ) 1 f (0) Along the corresponding line ] ] ]. p(α) p (0) + αs (0), the criterion function is ] f(α) f(p(α)). The minimum along this line follows from f! 0 α (0)
13 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A7.2 which yields the improved solution p (1) p(α (0) ) and so we got the minimizer in the first step. Conjugate Gradient Method Starting from the initial point p (0) 0.1, 1] we get the following iteration steps: ν p (v) f (v) exact line search 0 s (0) f (0) ] ] ] p(α) ] f(α) f! 0 α (0) 1 s (1) f (1) + f (1) 2 f (0) 2 s(0) ] ] ] p(α) ] f(α) 2 0 ] convergence in two steps 0 f! 0 α (1)
14 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A8.1 Minimization of a Quadratic Function Minimize the quadratic function f(p) 10p p 2 2. Quasi Newton Method Starting from the initial point p (0) 0.1, 1] we get the following iteration steps: ν p (v) f (v) exact line search 0 H (0) E ] ] ] s (0) H (0) f (0) ] p(α) ] f(α) f! 0 α (0) 1 ] ] δp p (1) p (0) ] γ f (1) f (0) ] δp γ, H (0) γ, γ H (0) γ
15 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A8.2 ν p (v) f (v) exact line search H (1) H (0) δpδp + H (1) ] s (1) H (1) f (1) ] p(α) ] f(α) f! 0 α (0) 2 0 ] convergence in two steps 0
16 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A9.1 Constrained Minimization of a Quadratic Function Minimize the quadratic function f(p) p p 2 2 subject to the constraint p 1 p 2 4 or written in the standard form h 1 (p) 0 Graphical Solution: For this simple problem, the solution can be found graphically in the design parameter space. The dimension of the design parameter space is. Plot the contour lines of the function f and the constraint h 1 and mark the optimal solution p.
17 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A9.2 Analytical Solution: The inequality constraint can be handled as an equality constraint by introducing the offset (slack) variable u 1 h 1 (p) 0 The Lagrange function for an optimization problem with constraints generally reads as It follows L(p, μ, u) f(p) μ_i(h i + u i 2 ) m i1 L(p, μ 1, u 1 ). The necessary condition of 1 st order for a minimum is L(p, μ 1, u 1 )! following four equations 0, which leads to the L p! ] ] L μ 1! L u 1! From the last equation we can distinguish two cases, either or.
18 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A9.3 Case 1: If μ 1 0, the constraint is not considered and therefore inactive. This is only true, if the constraint is fulfilled. It follows from L(p, μ 1, u 1 )! 0 p 1, p 2 and This last equation cannot be satisfied and subsequently the constraint is not fulfilled at the minimum of f at p ]and case 1 is not true. Case 2: If u 1 0, the distance from the limit of the constraint is 0 and at p we have an active constraint. The condition L(p, μ 1, u 1 )! 0 leaves us with three equations and it follows p 1, p 2, μ 1.
19 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A10.1 Constrained Minimization of a Nonlinear Function Minimize the nonlinear objective function f(p) (p p 2 ) 2 100p 1 subject to the equality constraint g(p) p p Solution by the Lagrange Newton Method: The Lagrange function for this optimization problem can be written as L(p, λ) λ. The Karush-Kuhn-Tucker (KKT) equations represent necessary conditions for an optimum. They can be written as L p! ] ] L λ! These nonlinear equations can be solved by the Newton-Raphson method. This yields the update equation of the Lagrange-Newton method 2 ( g(p) T p ) δp (v+1) δf(p) δp g(p) p ] p (v), λ (v) A(p,λ) λ (v+1) ] g(p) ] p (v) b(p)
20 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A10.2 With the objective and constraint function given above, the matrix A and the right hand side b follows as 30p A(p, λ) 200 2λ 2p 2, 0 ] b(p) 20p 1 3. ] Starting with p (0) 1 0 ] and λ (0) 0 the update equation reads as 30 δp (1) ] λ (1). ] 8 ]
21 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A10.3 From these three scalar equations, the updated optimization variables and the new Lagrange multiplier are given by p (1) p (0) + δp (1) ], λ (1) The next iteration yields p (2) λ (2) 2240, 2.6 ] ] ] which results in p (2) ] and λ (2) The following table summarizes the first 10 iterations of the Lagrange-Newton method: Iteration p 1 p 2 λ f g 0 1,0000E+00 0,0000E+00 0,0000E+00-9,9000E+01-8,0000E ,0000E+00-1,3000E+00-2,6000E+01 1,2044E+04 1,7690E ,7348E+00 6,3777E-01-3,1944E+02 3,0458E+03 5,3556E ,1496E+00-1,3357E-01 2,5548E+01 5,7946E+02 9,3751E ,8840E+00-2,8855E+00-1,1348E+02-2,6471E+02 7,6436E ,5959E+00-1,8490E+00-3,4899E+01-2,5859E+02 1,1574E ,5655E+00-1,5787E+00-1,0998E+01-2,5534E+02 7,3970E ,5667E+00-1,5533E+00-8,8899E+00-2,5478E+02 6,4691E ,5667E+00-1,5531E+00-8,8713E+00-2,5477E+02 4,6864E ,5667E+00-1,5531E+00-8,8713E+00-2,5477E+02 3,5527E ,5667E+00-1,5531E+00-8,8713E+00-2,5477E+02 1,0011E-15 Draw the points p (0) to p (4) and the final point p (10) in the contour plot on the next page. Another solution process was started with the initial point p (0) 1 following table summarizes the corresponding optimization history: 0] and λ (0) 0. The
22 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A10.4 Iteration p 1 p 2 λ f g 0-1,0000E+00 0,0000E+00 0,0000E+00 1,0100E+02-8,0000E ,0000E+00 1,3000E+00 7,4000E+01 1,3044E+04 1,7690E ,7693E+00-7,7022E-01-1,9290E+02 4,1289E+03 5,8005E ,2153E+00 2,8425E-01 8,7719E+01 1,2455E+03 1,4188E ,6628E+00 4,0384E+00-3,0435E+02 7,2870E+02 1,4399E ,3117E+00 2,4872E+00-8,3576E+01 3,8788E+02 2,5296E ,4700E+00 1,8314E+00-8,2745E+00 2,5753E+02 4,5508E ,4825E+00 1,6904E+00 8,1362E+00 2,5082E+02 2,0039E ,4838E+00 1,6826E+00 8,9300E+00 2,5064E+02 6,3143E ,4838E+00 1,6825E+00 8,9323E+00 2,5064E+02 4,5733E ,4838E+00 1,6825E+00 8,9323E+00 2,5064E+02-3,5527E-15 Draw the points p (0) to p (4) and the final point p (10) of this run in the contour plot, too. Contour plot of the constrained problem: As we can see, the solution depends on the starting point p (10). Visit the webpage and try to solve the problem with the stochastic particle swarm optimizer which is available there. What can you observe? What are the differences with respect to the Lagrange-Newton method?
23 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A11.1 Application: Optimization of a Double Pendulum An actively controlled double pendulum is described by the generalized coordinates y α β]. Without the torque actuator T, the dynamic behavior is described by the nonlinear equations of motion M(t, y) y + k(t, y, y ) q(t, y, y ) and the initial conditions are y(0) y 0, y (0) 0. Linearization of the equations of motion about the equilibrium position results in y(t) 0 + η(t), y (t) 0 + η (t), y (t) 0 + η (t) and the following system of linear equations where M(t) η (t) + P(t) η (t) + Q(t) η(t) h(t), M 2mL2 ml mgL 0 ml 2 2 ], P ], Q ml mgl ]. The generalized force vector q has to be considered with the control torque T(t) of the torque actuator h(t) T(t) 0 ] and the PD control law is T(t) k 1 α k 2 α
24 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A11.2 By an optimization process the control parameters k 1 and k 2 are to be optimized in such a way, that the vibrations of the pendulum, evaluated by the integral of y y within the first 10 s, are optimally damped. In order to ensure the stability of the system, both of the control parameters must not be negative. First we formulate the problem in standard form: design variables p ʼ ] kinetics initial conditions φ 0 0, φ 0 0. ] ] final condition H 1 0 criteria functions Ψ 1 Ψ 2 Ψ 3 optimization criterion inequality constraints f h 1 The constrained optimization task then reads as h 2 find, where and
25 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A11.3 The described optimization task is solved by a SQP method. The parameters of the double pendulum are m 1kg, L 0.3m, y ], and the initial design variable vector is p(0) ]. The optimization algorithm needs 10 gradient evaluations and 16 function evaluations in order to find the optimizer p ]. The optimization history shows the convergence after 10 iterations: The optimal dynamic behaviour shows a significantly improved damping in comparison to the passive and the initial design:
26 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A12.1 Multicriteria Optimization of a Truss For the presented truss the cross sections A 1 and A 2 of the two bars are to be designed. Simultaneously, the overall mass and the vertical displacement of the point P shall be minimized, while an external load F is applied at P. This is a typical problem of multicriteria optimization, since the overall mass corresponds to the costs of material and the vertical displacement is a measure for the function of the framework. Using the density ρ the overall mass is M For the computation of the vertical displacement the mass of the truss can be neglected in comparison to the external load. Using finite element modelling we get for the displacement y of the joint K y q where K E 2 2l A 2 A 2 ], y u A 2 2 2A 1 + A 2 v ], q 0 F ] which yields the vertical displacement v If A is a given minimal cross section of the bars, we get dimensionless design variables by normalization: p 1, p 2 - and the dimensionless optimization criteria f 1 M ρla, f 2 v Fl AE
27 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A12.2 If the possible cross section has a maximum of αa, then we get the vector optimization problem opt p P ] where P {p R 2 } Sketch the problem in the design space and in the criteria space for α 4. f(p)
28 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A12.3 From these sketches we get the Edgeworth Pareto optimal solutions P P { ], } For the numerical determination of the EP optimal solutions we generally formulate several scalar optimization problems. Formulate the following scalar optimization problems and determine the corresponding solution graphically. 1) Weighted Criteria Method (Weighting factors:w 1 w 2 1 2) min p P 2) Distance Method (Minimization of the Euklidean distance to the ideal solution f 0 ʼ ]) min p P
29 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A12.4 3) Hierarchical Method If the minimal vertical displacement is more important than the minimal overall mass of the truss, then the priorities are l 1 and l 2. In step 1, the minimization of the vertical displacement only leads to f 2 min p P 0 f 2(p) where P 0 In step 2, we try to additionally minimize the overall mass. For this, we have to allow some increase of the vertical displacement, e.g. ε 1 100%. The second optimization step then reads f 1 min p P 1 f 1(p) where P 1 { }
30 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A13.1 MATLAB basics Numerical Tools Optimization with MATLAB The program MATLAB is a commercially available program. It uses a numeric representation of vectors and matrices and allows a simple programming of matrix operations and provides many different methods to analyse systems. Many so-called Toolboxes are available to extend the functionality of MATLAB. The commands can be entered directly at the MATLAB prompt and executed immediately. Alternatively the commands can be summarized in a text file <name>.m. This socalled m-file is simply loaded and executed in MATLAB by typing <name> at the MATLAB prompt. helpdesk a ] b ]' A 1 2 4; ; ] starts online help defines a row vector defines a column vector defines matrix A transpose of matrix A A(2,3) A(:,3) A(1,:) zeros(n,m) ones(n,m) diag(1 2 3]) c A * b c A \ b inv(a) matrix element A23 3 rd column of matrix A 1 st row of matrix A defines a zero matrix defines a matrix of ones defines a diagonal matrix matrix-vector multiplication solves linear system inverse of a matrix plot(t,x, g ),hold on; EV, ew ] eig(a) matrix EV of eigenvectors and vector ew of eigenvalues of matrix A plot vector x over vector t For the numerical integration MATLAB provides two explicit single-step methods (ode23, ode45), one implicit single-step method (ode23s), an Adams-Bashfort- Moulton predictor corrector method (ode113) and an implicit numerical differentiation formula (ode15s). All methods have implemented a step size control. t,x] ode23('<name>',tspan,x0) simulation of a system of ordinary differential equations, given in the form dx/dt f(t,x). Outputs are the vector t of time steps and
31 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A13.2 the matrix X of state trajectories. The number of columns in X corresponds to the number of states and the number of rows to the number of time steps. Input is the <name> of a function specifying dx/dt, a (2x1) vector tspan t0, tend] contains the initial and the final time and the initial state vector is x0 1 ( t ) x ( t ) x ( t ) x ( t ) ]. x The function <name> must be written in a separate file <name.m> with the corresponding file name and must follow the syntax: function dx <name>(t, x)... dx... Example: Linearized double pendulum function dx doublependulum(t,x) M1 1.0; IZZ 0.0; GE 9.81; L 0.3; M 2.0*M1*L^2+IZZ M1*L^2; M1*L^2 M1*L^2+IZZ ]; Q 2*GE*M1*L 0; 0 GE*M1*L ]; A zeros(2) diag( 1 1]); -inv(m)*q zeros(2)]; dx A*x; MATLAB Optimization Toolbox The MATALAB Optimization Toolbox provides deterministic optimization algorithms for linear and non-nonlinear problems. The toolbox includes different solution methods for minimization problems, equations of nonlinear systems (fsolve, fzero) and least squares problems (curve fitting). Some important commands (described for medium scale optimization problems): fminsearch fminunc Finds minimum of unconstrained multivariable function using the derivative-free Simplex Search Method. Finds minimum of unconstrained multivariable function. The algorithm uses a Quasi-Newton method with a cubic line search procedure. This Quasi-Newton method uses the Broyden-Fletcher-Goldfarb-Shanno formula for updating the approximation of the Hessian matrix or the Davidson- Fletcher-Powell formula for approximating the inverse Hessian matrix.
32 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A13.3 fmincon Finds minimum of constrained nonlinear multivariable function. The algorithm uses a SQP method. During each iteration step an estimate of the Hessian of the Lagrangian is updated and a line search is performed. Example: Multicriteria Optimization of a truss structure using weighted criteria (A11) Main m-file truss.m for optimization: clear all; global w1 w2 % bounds of design variables alpha 3; A 1 0; 0 1; -1 0 ; 0-1]; b alpha alpha -1-1]'; % start values x0 1 1]; % options for optimization algorithm options optimset('largescale','off'); % weighting factors w1 0.3; w2 1-w1; % optimization x,fval] fmincon(@criteria,x0,a,b,],],],],],options) M-file criteria.m for function evaluation: function f criteria(x) global w1 w2 % calculation of optimization criteria % criteria f1 and f2 f1 x(1)+sqrt(2)*x(2); f2 1/x(1); % scalar optimization criteria f w1*f1+w2*f2; Example: Computation of the Pareto-Front of a truss structure (A11) The file truss.m is extended in order to compute the entire Pareto-Front. Thereby the weighted criteria method with variable weights is used. clear all; global w1 w2 alpha 3; % bounds of design variables A 1 0; 0 1; -1 0 ; 0-1]; b alpha alpha -1-1]'; % start values x0 1 1]; % options for optimization algorithm
33 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A13.4 options optimset('largescale','off'); for i1:51; % weighting factors w1 (i-1)/50; w2 1-w1; % optimization x(:,i),fval] fmincon(@criteria,x0,a,b,... ],],],],],options); % use previous results as next starting values x0 x(:,i); % criteria f1 and f2 f1(i) x(1,i)+sqrt(2)*x(2,i); f2(i) 1/x(1,i); end figure(1) subplot(1,2,1) plot(1 alpha],1 1],'b'), hold on; plot(1 1],1 alpha],'b'); plot(1 alpha],alpha alpha],'b'); plot(alpha alpha],1 alpha],'b'); plot(x(1,:),x(2,:),'r+'); axis(0.8 alpha alpha+0.2]) xlabel('p_1') ylabel('p_2') subplot(1,2,2) plot(f1,f2,'r+'), hold on; axis( ]) xlabel('f_1') ylabel('f_2') Figure 1: Pareto-Front of the truss structure. What problem do you see? MATLAB Genetic Algorithm and Direct Search Toolbox The MATLAB Genetic Algorithm and Direct Search Toolbox is a collection of stochastic optimization algorithms. This toolbox includes routines for solving optimization problems using Direct Search, Genetic Algorithm and Simulated Annealing.
34 Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 A13.5 Optimization with EXCEL For optimizing nonlinear problems EXCEL uses the Generalized Reduced Gradient (GRG2) algorithm developed by Leon Lasdon and Allan Waren. Other solvers can be installed, see e.g.
5 Handling Constraints
5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest
More informationOptimization Methods
Optimization Methods Decision making Examples: determining which ingredients and in what quantities to add to a mixture being made so that it will meet specifications on its composition allocating available
More informationEAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science
EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Multidimensional Unconstrained Optimization Suppose we have a function f() of more than one
More information4M020 Design tools. Algorithms for numerical optimization. L.F.P. Etman. Department of Mechanical Engineering Eindhoven University of Technology
4M020 Design tools Algorithms for numerical optimization L.F.P. Etman Department of Mechanical Engineering Eindhoven University of Technology Wednesday September 3, 2008 1 / 32 Outline 1 Problem formulation:
More informationAM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α
More informationAM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality
More informationLecture V. Numerical Optimization
Lecture V Numerical Optimization Gianluca Violante New York University Quantitative Macroeconomics G. Violante, Numerical Optimization p. 1 /19 Isomorphism I We describe minimization problems: to maximize
More informationISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints
ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints Instructor: Prof. Kevin Ross Scribe: Nitish John October 18, 2011 1 The Basic Goal The main idea is to transform a given constrained
More informationOptimization Methods
Optimization Methods Categorization of Optimization Problems Continuous Optimization Discrete Optimization Combinatorial Optimization Variational Optimization Common Optimization Concepts in Computer Vision
More informationScientific Computing: Optimization
Scientific Computing: Optimization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 March 8th, 2011 A. Donev (Courant Institute) Lecture
More informationNumerical Optimization: Basic Concepts and Algorithms
May 27th 2015 Numerical Optimization: Basic Concepts and Algorithms R. Duvigneau R. Duvigneau - Numerical Optimization: Basic Concepts and Algorithms 1 Outline Some basic concepts in optimization Some
More informationAlgorithms for Constrained Optimization
1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic
More informationReview for Exam 2 Ben Wang and Mark Styczynski
Review for Exam Ben Wang and Mark Styczynski This is a rough approximation of what we went over in the review session. This is actually more detailed in portions than what we went over. Also, please note
More information2.3 Linear Programming
2.3 Linear Programming Linear Programming (LP) is the term used to define a wide range of optimization problems in which the objective function is linear in the unknown variables and the constraints are
More informationOptimization and Root Finding. Kurt Hornik
Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding
More informationNumerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09
Numerical Optimization 1 Working Horse in Computer Vision Variational Methods Shape Analysis Machine Learning Markov Random Fields Geometry Common denominator: optimization problems 2 Overview of Methods
More informationNONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)
NONLINEAR PROGRAMMING (Hillier & Lieberman Introduction to Operations Research, 8 th edition) Nonlinear Programming g Linear programming has a fundamental role in OR. In linear programming all its functions
More informationStructural and Multidisciplinary Optimization. P. Duysinx and P. Tossings
Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be
More informationECS550NFB Introduction to Numerical Methods using Matlab Day 2
ECS550NFB Introduction to Numerical Methods using Matlab Day 2 Lukas Laffers lukas.laffers@umb.sk Department of Mathematics, University of Matej Bel June 9, 2015 Today Root-finding: find x that solves
More informationNonlinear Optimization: What s important?
Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global
More informationNumerical Methods for the Solution of Differential Equations
Numerical Methods for the Solution of Differential Equations Markus Grasmair Vienna, winter term 2011 2012 Analytical Solutions of Ordinary Differential Equations 1. Find the general solution of the differential
More informationInterior-Point Methods for Linear Optimization
Interior-Point Methods for Linear Optimization Robert M. Freund and Jorge Vera March, 204 c 204 Robert M. Freund and Jorge Vera. All rights reserved. Linear Optimization with a Logarithmic Barrier Function
More informationOptimization of Mechanical Systems
Prof. Dr.-Ing. Prof. E.h. P. Eberhard WT 18/19 M0 Optimization of Mechanical Systems Prof. Dr.-Ing. Prof. E.h. P. Eberhard (peter.eberhard@itm.uni-stuttgart.de) M.Sc. Elizaveta Shishova (elizaveta.shishova@itm.uni-stuttgart.de)
More informationIE 5531: Engineering Optimization I
IE 5531: Engineering Optimization I Lecture 12: Nonlinear optimization, continued Prof. John Gunnar Carlsson October 20, 2010 Prof. John Gunnar Carlsson IE 5531: Engineering Optimization I October 20,
More informationSMO Algorithms for Support Vector Machines without Bias Term
Institute of Automatic Control Laboratory for Control Systems and Process Automation Prof. Dr.-Ing. Dr. h. c. Rolf Isermann SMO Algorithms for Support Vector Machines without Bias Term Michael Vogt, 18-Jul-2002
More informationMultivariate Newton Minimanization
Multivariate Newton Minimanization Optymalizacja syntezy biosurfaktantu Rhamnolipid Rhamnolipids are naturally occuring glycolipid produced commercially by the Pseudomonas aeruginosa species of bacteria.
More informationNumerical Optimization. Review: Unconstrained Optimization
Numerical Optimization Finding the best feasible solution Edward P. Gatzke Department of Chemical Engineering University of South Carolina Ed Gatzke (USC CHE ) Numerical Optimization ECHE 589, Spring 2011
More informationNonlinear Programming (Hillier, Lieberman Chapter 13) CHEM-E7155 Production Planning and Control
Nonlinear Programming (Hillier, Lieberman Chapter 13) CHEM-E7155 Production Planning and Control 19/4/2012 Lecture content Problem formulation and sample examples (ch 13.1) Theoretical background Graphical
More informationMultidisciplinary System Design Optimization (MSDO)
Multidisciplinary System Design Optimization (MSDO) Numerical Optimization II Lecture 8 Karen Willcox 1 Massachusetts Institute of Technology - Prof. de Weck and Prof. Willcox Today s Topics Sequential
More informationNumerical Methods I Solving Nonlinear Equations
Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)
More informationConstrained optimization: direct methods (cont.)
Constrained optimization: direct methods (cont.) Jussi Hakanen Post-doctoral researcher jussi.hakanen@jyu.fi Direct methods Also known as methods of feasible directions Idea in a point x h, generate a
More informationThe roots are found with the following two statements. We have denoted the polynomial as p1, and the roots as roots_ p1.
Part II Lesson 10 Numerical Analysis Finding roots of a polynomial In MATLAB, a polynomial is expressed as a row vector of the form [an an 1 a2 a1 a0]. The elements ai of this vector are the coefficients
More informationReal-time Constrained Nonlinear Optimization for Maximum Power Take-off of a Wave Energy Converter
Real-time Constrained Nonlinear Optimization for Maximum Power Take-off of a Wave Energy Converter Thomas Bewley 23 May 2014 Southern California Optimization Day Summary 1 Introduction 2 Nonlinear Model
More informationUnconstrained Multivariate Optimization
Unconstrained Multivariate Optimization Multivariate optimization means optimization of a scalar function of a several variables: and has the general form: y = () min ( ) where () is a nonlinear scalar-valued
More informationWritten Examination
Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes
More informationNonlinearOptimization
1/35 NonlinearOptimization Pavel Kordík Department of Computer Systems Faculty of Information Technology Czech Technical University in Prague Jiří Kašpar, Pavel Tvrdík, 2011 Unconstrained nonlinear optimization,
More informationQuasi-Newton Methods
Newton s Method Pros and Cons Quasi-Newton Methods MA 348 Kurt Bryan Newton s method has some very nice properties: It s extremely fast, at least once it gets near the minimum, and with the simple modifications
More informationNumerical Optimization of Partial Differential Equations
Numerical Optimization of Partial Differential Equations Part I: basic optimization concepts in R n Bartosz Protas Department of Mathematics & Statistics McMaster University, Hamilton, Ontario, Canada
More informationPart 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)
Part 4: Active-set methods for linearly constrained optimization Nick Gould RAL fx subject to Ax b Part C course on continuoue optimization LINEARLY CONSTRAINED MINIMIZATION fx subject to Ax { } b where
More informationMotivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:
CDS270 Maryam Fazel Lecture 2 Topics from Optimization and Duality Motivation network utility maximization (NUM) problem: consider a network with S sources (users), each sending one flow at rate x s, through
More informationCHAPTER 2: QUADRATIC PROGRAMMING
CHAPTER 2: QUADRATIC PROGRAMMING Overview Quadratic programming (QP) problems are characterized by objective functions that are quadratic in the design variables, and linear constraints. In this sense,
More informationStatic unconstrained optimization
Static unconstrained optimization 2 In unconstrained optimization an objective function is minimized without any additional restriction on the decision variables, i.e. min f(x) x X ad (2.) with X ad R
More informationAppendix A Taylor Approximations and Definite Matrices
Appendix A Taylor Approximations and Definite Matrices Taylor approximations provide an easy way to approximate a function as a polynomial, using the derivatives of the function. We know, from elementary
More informationNonlinear Optimization for Optimal Control
Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]
More informationA Trust-region-based Sequential Quadratic Programming Algorithm
Downloaded from orbit.dtu.dk on: Oct 19, 2018 A Trust-region-based Sequential Quadratic Programming Algorithm Henriksen, Lars Christian; Poulsen, Niels Kjølstad Publication date: 2010 Document Version
More informationLecture 18: Optimization Programming
Fall, 2016 Outline Unconstrained Optimization 1 Unconstrained Optimization 2 Equality-constrained Optimization Inequality-constrained Optimization Mixture-constrained Optimization 3 Quadratic Programming
More informationWhat s New in Active-Set Methods for Nonlinear Optimization?
What s New in Active-Set Methods for Nonlinear Optimization? Philip E. Gill Advances in Numerical Computation, Manchester University, July 5, 2011 A Workshop in Honor of Sven Hammarling UCSD Center for
More informationGradient Descent. Dr. Xiaowei Huang
Gradient Descent Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Three machine learning algorithms: decision tree learning k-nn linear regression only optimization objectives are discussed,
More informationNeural Network Training
Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification
More informationNumerical optimization
Numerical optimization Lecture 4 Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Numerical geometry of non-rigid shapes Stanford University, Winter 2009 2 Longest Slowest Shortest Minimal Maximal
More informationMATH2070 Optimisation
MATH2070 Optimisation Nonlinear optimisation with constraints Semester 2, 2012 Lecturer: I.W. Guo Lecture slides courtesy of J.R. Wishart Review The full nonlinear optimisation problem with equality constraints
More informationAn Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84
An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A
More informationMath 409/509 (Spring 2011)
Math 409/509 (Spring 2011) Instructor: Emre Mengi Study Guide for Homework 2 This homework concerns the root-finding problem and line-search algorithms for unconstrained optimization. Please don t hesitate
More informationMath Ordinary Differential Equations
Math 411 - Ordinary Differential Equations Review Notes - 1 1 - Basic Theory A first order ordinary differential equation has the form x = f(t, x) (11) Here x = dx/dt Given an initial data x(t 0 ) = x
More informationInterior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems
AMSC 607 / CMSC 764 Advanced Numerical Optimization Fall 2008 UNIT 3: Constrained Optimization PART 4: Introduction to Interior Point Methods Dianne P. O Leary c 2008 Interior Point Methods We ll discuss
More informationNewton s Method. Ryan Tibshirani Convex Optimization /36-725
Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x
More informationPrimal-Dual Interior-Point Methods for Linear Programming based on Newton s Method
Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method Robert M. Freund March, 2004 2004 Massachusetts Institute of Technology. The Problem The logarithmic barrier approach
More informationOptimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23
Optimization: Nonlinear Optimization without Constraints Nonlinear Optimization without Constraints 1 / 23 Nonlinear optimization without constraints Unconstrained minimization min x f(x) where f(x) is
More informationNumerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems
1 Numerical optimization Alexander & Michael Bronstein, 2006-2009 Michael Bronstein, 2010 tosca.cs.technion.ac.il/book Numerical optimization 048921 Advanced topics in vision Processing and Analysis of
More informationIn view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written
11.8 Inequality Constraints 341 Because by assumption x is a regular point and L x is positive definite on M, it follows that this matrix is nonsingular (see Exercise 11). Thus, by the Implicit Function
More informationOptimization Methods for Circuit Design
Technische Universität München Department of Electrical Engineering and Information Technology Institute for Electronic Design Automation Optimization Methods for Circuit Design Compendium H. Graeb Version
More informationNumerical Optimization Algorithms
Numerical Optimization Algorithms 1. Overview. Calculus of Variations 3. Linearized Supersonic Flow 4. Steepest Descent 5. Smoothed Steepest Descent Overview 1 Two Main Categories of Optimization Algorithms
More informationOutline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St
Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic
More informationCEE Computer Applications. Mathematical Programming (LP) and Excel Solver
CEE 3804 - Computer Applications Mathematical Programming (LP) and Excel Solver Dr. Antonio A. Trani Professor of Civil and Environmental Engineering Virginia Polytechnic Institute and State University
More informationSF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren
SF2822 Applied Nonlinear Optimization Lecture 9: Sequential quadratic programming Anders Forsgren SF2822 Applied Nonlinear Optimization, KTH / 24 Lecture 9, 207/208 Preparatory question. Try to solve theory
More informationMATH 4211/6211 Optimization Quasi-Newton Method
MATH 4211/6211 Optimization Quasi-Newton Method Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Quasi-Newton Method Motivation:
More informationConstrained Optimization
1 / 22 Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 30, 2015 2 / 22 1. Equality constraints only 1.1 Reduced gradient 1.2 Lagrange
More informationLecture 3. Optimization Problems and Iterative Algorithms
Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex
More information5 Quasi-Newton Methods
Unconstrained Convex Optimization 26 5 Quasi-Newton Methods If the Hessian is unavailable... Notation: H = Hessian matrix. B is the approximation of H. C is the approximation of H 1. Problem: Solve min
More informationCHAPTER 12 TIME DOMAIN: MODAL STATE SPACE FORM
CHAPTER 1 TIME DOMAIN: MODAL STATE SPACE FORM 1.1 Introduction In Chapter 7 we derived the equations of motion in modal form for the system in Figure 1.1. In this chapter we will convert the modal form
More informationOptimization Concepts and Applications in Engineering
Optimization Concepts and Applications in Engineering Ashok D. Belegundu, Ph.D. Department of Mechanical Engineering The Pennsylvania State University University Park, Pennsylvania Tirupathi R. Chandrupatia,
More informationMATH 3795 Lecture 13. Numerical Solution of Nonlinear Equations in R N.
MATH 3795 Lecture 13. Numerical Solution of Nonlinear Equations in R N. Dmitriy Leykekhman Fall 2008 Goals Learn about different methods for the solution of F (x) = 0, their advantages and disadvantages.
More informationMATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year
MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2 1 Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year 2013-14 OUTLINE OF WEEK 8 topics: quadratic optimisation, least squares,
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization 1 Origins Instructor: Michael Saunders Spring 2015 Notes 9: Augmented Lagrangian Methods
More informationSTATE VARIABLE (SV) SYSTEMS
Copyright F.L. Lewis 999 All rights reserved Updated:Tuesday, August 05, 008 STATE VARIABLE (SV) SYSTEMS A natural description for dynamical systems is the nonlinear state-space or state variable (SV)
More informationDetermination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study
International Journal of Mathematics And Its Applications Vol.2 No.4 (2014), pp.47-56. ISSN: 2347-1557(online) Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms:
More informationAlgorithms for nonlinear programming problems II
Algorithms for nonlinear programming problems II Martin Branda Charles University Faculty of Mathematics and Physics Department of Probability and Mathematical Statistics Computational Aspects of Optimization
More informationContents. Preface. 1 Introduction Optimization view on mathematical models NLP models, black-box versus explicit expression 3
Contents Preface ix 1 Introduction 1 1.1 Optimization view on mathematical models 1 1.2 NLP models, black-box versus explicit expression 3 2 Mathematical modeling, cases 7 2.1 Introduction 7 2.2 Enclosing
More informationAlgorithms for constrained local optimization
Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained
More informationMODERN CONTROL DESIGN
CHAPTER 8 MODERN CONTROL DESIGN The classical design techniques of Chapters 6 and 7 are based on the root-locus and frequency response that utilize only the plant output for feedback with a dynamic controller
More informationNonlinear Programming
Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week
More informationComputational Finance
Department of Mathematics at University of California, San Diego Computational Finance Optimization Techniques [Lecture 2] Michael Holst January 9, 2017 Contents 1 Optimization Techniques 3 1.1 Examples
More informationConstrained Optimization and Lagrangian Duality
CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may
More informationOptimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29
Optimization Yuh-Jye Lee Data Science and Machine Intelligence Lab National Chiao Tung University March 21, 2017 1 / 29 You Have Learned (Unconstrained) Optimization in Your High School Let f (x) = ax
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428
More informationOrdinary differential equations - Initial value problems
Education has produced a vast population able to read but unable to distinguish what is worth reading. G.M. TREVELYAN Chapter 6 Ordinary differential equations - Initial value problems In this chapter
More informationMathematical Economics. Lecture Notes (in extracts)
Prof. Dr. Frank Werner Faculty of Mathematics Institute of Mathematical Optimization (IMO) http://math.uni-magdeburg.de/ werner/math-ec-new.html Mathematical Economics Lecture Notes (in extracts) Winter
More informationFALL 2018 MATH 4211/6211 Optimization Homework 4
FALL 2018 MATH 4211/6211 Optimization Homework 4 This homework assignment is open to textbook, reference books, slides, and online resources, excluding any direct solution to the problem (such as solution
More informationApplied Computational Economics Workshop. Part 3: Nonlinear Equations
Applied Computational Economics Workshop Part 3: Nonlinear Equations 1 Overview Introduction Function iteration Newton s method Quasi-Newton methods Practical example Practical issues 2 Introduction Nonlinear
More informationChapter 2. Optimization. Gradients, convexity, and ALS
Chapter 2 Optimization Gradients, convexity, and ALS Contents Background Gradient descent Stochastic gradient descent Newton s method Alternating least squares KKT conditions 2 Motivation We can solve
More informationOptimization Problems with Constraints - introduction to theory, numerical Methods and applications
Optimization Problems with Constraints - introduction to theory, numerical Methods and applications Dr. Abebe Geletu Ilmenau University of Technology Department of Simulation and Optimal Processes (SOP)
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table
More informationOptimisation in Higher Dimensions
CHAPTER 6 Optimisation in Higher Dimensions Beyond optimisation in 1D, we will study two directions. First, the equivalent in nth dimension, x R n such that f(x ) f(x) for all x R n. Second, constrained
More informationminimize x subject to (x 2)(x 4) u,
Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for
More informationConstrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.
Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization
More informationWHY DUALITY? Gradient descent Newton s method Quasi-newton Conjugate gradients. No constraints. Non-differentiable ???? Constrained problems? ????
DUALITY WHY DUALITY? No constraints f(x) Non-differentiable f(x) Gradient descent Newton s method Quasi-newton Conjugate gradients etc???? Constrained problems? f(x) subject to g(x) apple 0???? h(x) =0
More informationNumerical Methods for Embedded Optimization and Optimal Control. Exercises
Summer Course Numerical Methods for Embedded Optimization and Optimal Control Exercises Moritz Diehl, Daniel Axehill and Lars Eriksson June 2011 Introduction This collection of exercises is intended to
More informationReview of Optimization Basics
Review of Optimization Basics. Introduction Electricity markets throughout the US are said to have a two-settlement structure. The reason for this is that the structure includes two different markets:
More informationSurvey of NLP Algorithms. L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA
Survey of NLP Algorithms L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA NLP Algorithms - Outline Problem and Goals KKT Conditions and Variable Classification Handling
More informationCE 191: Civil & Environmental Engineering Systems Analysis. LEC 17 : Final Review
CE 191: Civil & Environmental Engineering Systems Analysis LEC 17 : Final Review Professor Scott Moura Civil & Environmental Engineering University of California, Berkeley Fall 2014 Prof. Moura UC Berkeley
More information