Conjugate Gradient Methods

Size: px
Start display at page:

Download "Conjugate Gradient Methods"

Transcription

1 Conjugate Gradient Methods Now we consider optimization in n-dimensional design space. The objective is to minimize function f(x), where the components of x are the n independent design variables. The basic strategy is 1. choose a point x in the design space. 2. loop with i = 1, 2, 3,... choose a vector v i minimize f(x) along the line through x i 1 in the direction of v i. let the minimum point be x i. If x i x i 1 < ε exit loop 3. end loop The minimization along a line can be accomplished with any one-dimensional optimization algorithm (such as the golden section search). The very important question is how to choose the vectors v i. Conjugate Directions Consider the quadratic function f(x) = c i Differentiation with respect to x i yields which can be written in vector notation as b i x i + 1 A ij x i x j = c b T x xt Ax. (1) i f x i = b i + j j A ij x j, (2) f = b + Ax, (3) where f is the gradient of f. Now consider the change in the gradient as we move from point x in the direction of a vector u. The motion takes place along the line x = x + su, (4) 1

2 2 where s is the distance moved. Substitution into Eq. (3) yields the expression for the gradient at x f x +su = b + A(x + su) = f x + sau. (5) Note that the change in the gradient is sau. If this change is perpendicular to a vector v, then v T Au =, (6) and the directions of u and v are said to be mutually conjugate (non-interfering). The implication is that once we have minimized f(x) in the direction of v, we can move along u without ruining the previous minimization. For a quadratic function of n independent variables it is possible to construct n mutually conjugate directions. Therefore, it would take exactly n line minimizations along these directions to reach the minimum point. If f(x) is not a quadratic function, Eq. (1) can be treated as a local approximation of the merit function, obtained by truncating the Taylor series expansion of f(x) about x f(x) f(x ) + f(x )(x x ) (x x ) T H(x x ). (7) Now the conjugate directions based on the quadratic form are only approximations, valid in the close vicinity of x. Consequently, it would take several cycles of n line minimizations to reach the optimal point. The various conjugate gradient methods use different techniques for constructing conjugate directions. The so-called zero-order methods work with f(x) only, whereas the first-order methods utilize both f(x) and f. The first-order methods are computationally more efficient, but the input of f can be very tedious, or it may be not available at all. Powell s Method Powell s method is a zero-order method, requiring the evaluation of f(x) only. involves n design variables, the basic algorithm is 1. Choose a point x in the design space. If the problem 2. Choose the starting vectors v i, i = 1, 2,..., n (the usual choice is v i = e i, where e i is the unit vector in the x i -coordinate direction). 3. Start cycle do i = 1, 2,..., n Minimize f(x) along the line through x i 1 in the direction of v i. Let the minimum point be x i. end do v n+1 x x n Minimize f(x) along the line through x in the direction of v n+1. Let the minimum point be x n+1. if x n+1 x < ε exit loop do i = 1, 2,..., n v i v i+1 (v 1 is discarded, the other vectors are reused). end do

3 3 4. End cycle Powell demonstrated that the vectors v n+1 produced in successive cycles are mutually conjugate, so that the minimum point of a quadratic surface is reached in exactly n cycles. In practice, the merit function is seldom quadratic, but as long as it can be approximated locally by Eq. (1), Powell s method will work. Of course, it takes more than n cycles to arrive at the minimum of a non-quadratic function. Example 1 Find the minimum of the function f = 1(y x 2 ) 2 + (1 x) 2 using Powell s method starting at the point (-1,1). This function has an interesting topology. There is a hump between the starting and minimum points. The minimum value of f occurs at the point (1,1). #!/usr/bin/python from powell import * from numpy import array def F(x): return 1.*(x[1] - x[]**2)**2 + (1 - x[])**2 xstart = array([-1., 1.]) xmin,niter = powell(f,xstart) print "x =",xmin print "F(x) =", F(xMin) print "Number of cycles =", niter raw_input ("") The minimum point is obtained after 12 cycles. x = [ 1. 1.] F(x) = e-29 Number of cycles = 12 Example 2 Use Powell s method to determine the smallest distance from the point (5,8) to the curve xy = 5. This is a constrained optimization problem: minimize f(x, y) = (x 5) 2 + (y 8) 2 (the square of the distances) subject to the equality constraint xy 5 =. The following program uses Powell s method with penalty function. #!/usr/bin/python from powell import * from numpy import array from math import sqrt def F(x):

4 4 lam = 1. # Penalty multiplier c = x[]*x[1] - 5. # Constraint equation return distsq(x) + lam*c**2 # Penalized merit function def distsq(x): return (x[] - 5)**2 + (x[1] - 8)**2 xstart = array([1.,5.]) x,numiter = powell(f,xstart,.1) print "Intersection point =", x print "Minimum distance =", sqrt(distsq(x)) print "xy =", x[]*x[1] print "Number of cycles =", numiter raw_input ("") The value of the penalty function multiplier µ can have profound effect on the result. With µ = 1 we obtain the following result: Intersection point = [ ] Minimum distance = xy = Number of cycles = 5 The small value of µ favored speed of convergence over accuracy. Because the violation of the constraint xy = 5 is clearly unacceptable, we run the script again with µ = 1 and change the starting point to ( ), the end point of the first run. The results shown next are now acceptable: Intersection point = [ ] Minimum distance = xy = Number of cycles = 5 A large µ often causes the algorithm to hang up, so that it is generally wise to start with a small µ. Example 3 The displacement formulation of the truss shown in Figure 1 results in the following system of equations for the joint displacements u E 2 2L 2 2A 2 + A 3 A 3 A 3 A 3 A 3 A 3 A 3 A 3 2 2A 1 + A 3 u 1 u 2 u 3 = P, (8) where E is the modulus of elasticity, and A i is the cross-sectional area of ith member. Use Powell s method to minimize the structural volume of the truss while keeping the displacement u 2 below a given value δ.

5 5 u 3 1 L L 3 u 2 2 P u 1 Figure 1: 1 Introducing the dimensionless variables v i = u i δ, the system of equations becomes 2 2x x 3 x 3 x 3 2 x 3 x 3 x 3 2 x 3 x 3 2 2x 1 + x 3 The structural volume to be minimized is x i = Eδ P L A i (9) v 1 v 2 v 3 = 1. (1) V = L(A 1 + A 2 + 2A 3 ) = P L2 Eδ (x 1 + x 2 + 2x 3 ) (11) In addition to the displacement constraint u 2 δ, we should also prevent the cross-sectional areas from becoming negative by applying the constraints A i. Thus, the optimization problem becomes minimize f = x 1 + x 2 + 2x 3 with the inequality constraints v 2 1, x i, i = 1, 2, 3. (12) Note that in order to obtain v 2 we must solve system (1). #!/usr/bin/python from powell import * from numpy import array from math import sqrt from gausselimin import * def F(x): global v, weight mu = 1. c = 2.*sqrt(2.)

6 6 A = array([[c*x[1] + x[2], -x[2], x[2]], \ [-x[2], x[2], -x[2]], \ [ x[2], -x[2], c*x[] + x[2]]])/c b = array([., -1.,.]) v = gausselimin(a,b) weight = x[] + x[1] + sqrt(2.)*x[2] penalty = max(.,abs(v[1]) - 1.)**2 \ + max(.,-x[])**2 \ + max(.,-x[1])**2 \ + max(.,-x[2])**2 return weight + penalty*mu xstart = array([1., 1., 1.]) x,numiter = powell(f,xstart) print "x = ", x print "v = ", v print "Relative weight F = ", weight print "Number of cycles = ", numiter raw_input ("") The first run of the program starts with x = [ ] T and uses µ = 1 for the penalty multiplier. The results are x = [ ] v = [ ] Relative weight F = Number of cycles = 1 Because the magnitude of v 2 is excessive, the penalty multiplier is increased to 1, and the program runs again using the output x from the last run as the input. x = [ ] v = [ ] Relative weight F = Number of cycles = 11 Now v 2 is much closer to the constraint value.

MVE165/MMG631 Linear and integer optimization with applications Lecture 13 Overview of nonlinear programming. Ann-Brith Strömberg

MVE165/MMG631 Linear and integer optimization with applications Lecture 13 Overview of nonlinear programming. Ann-Brith Strömberg MVE165/MMG631 Overview of nonlinear programming Ann-Brith Strömberg 2015 05 21 Areas of applications, examples (Ch. 9.1) Structural optimization Design of aircraft, ships, bridges, etc Decide on the material

More information

MS 2001: Test 1 B Solutions

MS 2001: Test 1 B Solutions MS 2001: Test 1 B Solutions Name: Student Number: Answer all questions. Marks may be lost if necessary work is not clearly shown. Remarks by me in italics and would not be required in a test - J.P. Question

More information

Constrained optimization: direct methods (cont.)

Constrained optimization: direct methods (cont.) Constrained optimization: direct methods (cont.) Jussi Hakanen Post-doctoral researcher jussi.hakanen@jyu.fi Direct methods Also known as methods of feasible directions Idea in a point x h, generate a

More information

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23 Optimization: Nonlinear Optimization without Constraints Nonlinear Optimization without Constraints 1 / 23 Nonlinear optimization without constraints Unconstrained minimization min x f(x) where f(x) is

More information

Functions of Several Variables

Functions of Several Variables Functions of Several Variables The Unconstrained Minimization Problem where In n dimensions the unconstrained problem is stated as f() x variables. minimize f()x x, is a scalar objective function of vector

More information

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL) Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization Nick Gould (RAL) x IR n f(x) subject to c(x) = Part C course on continuoue optimization CONSTRAINED MINIMIZATION x

More information

Functions and Equations

Functions and Equations Canadian Mathematics Competition An activity of the Centre for Education in Mathematics and Computing, University of Waterloo, Waterloo, Ontario Euclid eworkshop # Functions and Equations c 006 CANADIAN

More information

Department of Mathematics, University of Wisconsin-Madison Math 114 Worksheet Sections 3.1, 3.3, and 3.5

Department of Mathematics, University of Wisconsin-Madison Math 114 Worksheet Sections 3.1, 3.3, and 3.5 Department of Mathematics, University of Wisconsin-Madison Math 11 Worksheet Sections 3.1, 3.3, and 3.5 1. For f(x) = 5x + (a) Determine the slope and the y-intercept. f(x) = 5x + is of the form y = mx

More information

Math 212-Lecture Interior critical points of functions of two variables

Math 212-Lecture Interior critical points of functions of two variables Math 212-Lecture 24 13.10. Interior critical points of functions of two variables Previously, we have concluded that if f has derivatives, all interior local min or local max should be critical points.

More information

Math 0320 Final Exam Review

Math 0320 Final Exam Review Math 0320 Final Exam Review SHORT ANSWER. Write the word or phrase that best completes each statement or answers the question. Factor out the GCF using the Distributive Property. 1) 6x 3 + 9x 1) Objective:

More information

Math M111: Lecture Notes For Chapter 3

Math M111: Lecture Notes For Chapter 3 Section 3.1: Math M111: Lecture Notes For Chapter 3 Note: Make sure you already printed the graphing papers Plotting Points, Quadrant s signs, x-intercepts and y-intercepts Example 1: Plot the following

More information

Nonlinear Optimization

Nonlinear Optimization Nonlinear Optimization (Com S 477/577 Notes) Yan-Bin Jia Nov 7, 2017 1 Introduction Given a single function f that depends on one or more independent variable, we want to find the values of those variables

More information

1 Solving Algebraic Equations

1 Solving Algebraic Equations Arkansas Tech University MATH 1203: Trigonometry Dr. Marcel B. Finan 1 Solving Algebraic Equations This section illustrates the processes of solving linear and quadratic equations. The Geometry of Real

More information

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema Chapter 7 Extremal Problems No matter in theoretical context or in applications many problems can be formulated as problems of finding the maximum or minimum of a function. Whenever this is the case, advanced

More information

AS PURE MATHS REVISION NOTES

AS PURE MATHS REVISION NOTES AS PURE MATHS REVISION NOTES 1 SURDS A root such as 3 that cannot be written exactly as a fraction is IRRATIONAL An expression that involves irrational roots is in SURD FORM e.g. 2 3 3 + 2 and 3-2 are

More information

Quartic Equation. By CH vd Westhuizen A unique Solution assuming Complex roots. Ax^4 + Bx^3 + Cx^2 + Dx + E = 0

Quartic Equation. By CH vd Westhuizen A unique Solution assuming Complex roots. Ax^4 + Bx^3 + Cx^2 + Dx + E = 0 Quartic Equation By CH vd Westhuizen A unique Solution assuming Complex roots The general Quartic is given by Ax^4 + Bx^3 + Cx^ + Dx + E = 0 As in the third order polynomial we are first going to reduce

More information

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

Algebra II Vocabulary Alphabetical Listing. Absolute Maximum: The highest point over the entire domain of a function or relation.

Algebra II Vocabulary Alphabetical Listing. Absolute Maximum: The highest point over the entire domain of a function or relation. Algebra II Vocabulary Alphabetical Listing Absolute Maximum: The highest point over the entire domain of a function or relation. Absolute Minimum: The lowest point over the entire domain of a function

More information

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be

More information

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method. Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization

More information

Elements of Continuum Elasticity. David M. Parks Mechanics and Materials II February 25, 2004

Elements of Continuum Elasticity. David M. Parks Mechanics and Materials II February 25, 2004 Elements of Continuum Elasticity David M. Parks Mechanics and Materials II 2.002 February 25, 2004 Solid Mechanics in 3 Dimensions: stress/equilibrium, strain/displacement, and intro to linear elastic

More information

Penalty and Barrier Methods General classical constrained minimization problem minimize f(x) subject to g(x) 0 h(x) =0 Penalty methods are motivated by the desire to use unconstrained optimization techniques

More information

Multivariate Newton Minimanization

Multivariate Newton Minimanization Multivariate Newton Minimanization Optymalizacja syntezy biosurfaktantu Rhamnolipid Rhamnolipids are naturally occuring glycolipid produced commercially by the Pseudomonas aeruginosa species of bacteria.

More information

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 11 and

More information

Optimality Conditions for Constrained Optimization

Optimality Conditions for Constrained Optimization 72 CHAPTER 7 Optimality Conditions for Constrained Optimization 1. First Order Conditions In this section we consider first order optimality conditions for the constrained problem P : minimize f 0 (x)

More information

Optimality Conditions

Optimality Conditions Chapter 2 Optimality Conditions 2.1 Global and Local Minima for Unconstrained Problems When a minimization problem does not have any constraints, the problem is to find the minimum of the objective function.

More information

Lecture 3. Optimization Problems and Iterative Algorithms

Lecture 3. Optimization Problems and Iterative Algorithms Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex

More information

Maths Higher Prelim Content

Maths Higher Prelim Content Maths Higher Prelim Content Straight Line Gradient of a line A(x 1, y 1 ), B(x 2, y 2 ), Gradient of AB m AB = y 2 y1 x 2 x 1 m = tanθ where θ is the angle the line makes with the positive direction of

More information

CALCULUS: Math 21C, Fall 2010 Final Exam: Solutions. 1. [25 pts] Do the following series converge or diverge? State clearly which test you use.

CALCULUS: Math 21C, Fall 2010 Final Exam: Solutions. 1. [25 pts] Do the following series converge or diverge? State clearly which test you use. CALCULUS: Math 2C, Fall 200 Final Exam: Solutions. [25 pts] Do the following series converge or diverge? State clearly which test you use. (a) (d) n(n + ) ( ) cos n n= n= (e) (b) n= n= [ cos ( ) n n (c)

More information

11. Iteration: The while-loop

11. Iteration: The while-loop 11. Iteration: The while-loop Topics: Open-Ended repetition the while statement Example 1: The sqrt Problem Example 2: The UpDown Sequence Example 3. The Fibonacci Sequence Open-Ended Iteration So far,

More information

Math 1302 Notes 2. How many solutions? What type of solution in the real number system? What kind of equation is it?

Math 1302 Notes 2. How many solutions? What type of solution in the real number system? What kind of equation is it? Math 1302 Notes 2 We know that x 2 + 4 = 0 has How many solutions? What type of solution in the real number system? What kind of equation is it? What happens if we enlarge our current system? Remember

More information

ALGEBRAIC GEOMETRY HOMEWORK 3

ALGEBRAIC GEOMETRY HOMEWORK 3 ALGEBRAIC GEOMETRY HOMEWORK 3 (1) Consider the curve Y 2 = X 2 (X + 1). (a) Sketch the curve. (b) Determine the singular point P on C. (c) For all lines through P, determine the intersection multiplicity

More information

1 Computing with constraints

1 Computing with constraints Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)

More information

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study International Journal of Mathematics And Its Applications Vol.2 No.4 (2014), pp.47-56. ISSN: 2347-1557(online) Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms:

More information

Optimization Theory. Lectures 4-6

Optimization Theory. Lectures 4-6 Optimization Theory Lectures 4-6 Unconstrained Maximization Problem: Maximize a function f:ú n 6 ú within a set A f ú n. Typically, A is ú n, or the non-negative orthant {x0ú n x$0} Existence of a maximum:

More information

CE 191: Civil and Environmental Engineering Systems Analysis. LEC 05 : Optimality Conditions

CE 191: Civil and Environmental Engineering Systems Analysis. LEC 05 : Optimality Conditions CE 191: Civil and Environmental Engineering Systems Analysis LEC : Optimality Conditions Professor Scott Moura Civil & Environmental Engineering University of California, Berkeley Fall 214 Prof. Moura

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted

More information

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST) Convex Functions Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Definition convex function Examples

More information

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly. C PROPERTIES OF MATRICES 697 to whether the permutation i 1 i 2 i N is even or odd, respectively Note that I =1 Thus, for a 2 2 matrix, the determinant takes the form A = a 11 a 12 = a a 21 a 11 a 22 a

More information

Friday, 2 November 12. Vectors

Friday, 2 November 12. Vectors Vectors Scalars We often employ a single number to represent quantities that we use in our daily lives such as weight, height etc. The magnitude of this number depends on our age and whether we use metric

More information

Twitter: @Owen134866 www.mathsfreeresourcelibrary.com Prior Knowledge Check 1) Simplify: a) 3x 2 5x 5 b) 5x3 y 2 15x 7 2) Factorise: a) x 2 2x 24 b) 3x 2 17x + 20 15x 2 y 3 3) Use long division to calculate:

More information

Lecture 35 Minimization and maximization of functions. Powell s method in multidimensions Conjugate gradient method. Annealing methods.

Lecture 35 Minimization and maximization of functions. Powell s method in multidimensions Conjugate gradient method. Annealing methods. Lecture 35 Minimization and maximization of functions Powell s method in multidimensions Conjugate gradient method. Annealing methods. We know how to minimize functions in one dimension. If we start at

More information

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A

More information

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained

More information

Error, Accuracy and Convergence

Error, Accuracy and Convergence Error, Accuracy and Convergence Error in Numerical Methods i Every result we compute in Numerical Methods is inaccurate. What is our model of that error? Approximate Result = True Value + Error. x = x

More information

Christmas Calculated Colouring - C1

Christmas Calculated Colouring - C1 Christmas Calculated Colouring - C Tom Bennison December 20, 205 Introduction Each question identifies a region or regions on the picture Work out the answer and use the key to work out which colour to

More information

TMA 4180 Optimeringsteori THE CONJUGATE GRADIENT METHOD

TMA 4180 Optimeringsteori THE CONJUGATE GRADIENT METHOD INTRODUCTION TMA 48 Optimeringsteori THE CONJUGATE GRADIENT METHOD H. E. Krogstad, IMF, Spring 28 This note summarizes main points in the numerical analysis of the Conjugate Gradient (CG) method. Most

More information

King Fahd University of Petroleum and Minerals Prep-Year Math Program Math Term 161 Recitation (R1, R2)

King Fahd University of Petroleum and Minerals Prep-Year Math Program Math Term 161 Recitation (R1, R2) Math 001 - Term 161 Recitation (R1, R) Question 1: How many rational and irrational numbers are possible between 0 and 1? (a) 1 (b) Finite (c) 0 (d) Infinite (e) Question : A will contain how many elements

More information

SOLUTIONS FOR PROBLEMS 1-30

SOLUTIONS FOR PROBLEMS 1-30 . Answer: 5 Evaluate x x + 9 for x SOLUTIONS FOR PROBLEMS - 0 When substituting x in x be sure to do the exponent before the multiplication by to get (). + 9 5 + When multiplying ( ) so that ( 7) ( ).

More information

Scientific Computing: Optimization

Scientific Computing: Optimization Scientific Computing: Optimization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 March 8th, 2011 A. Donev (Courant Institute) Lecture

More information

Linear Regression. S. Sumitra

Linear Regression. S. Sumitra Linear Regression S Sumitra Notations: x i : ith data point; x T : transpose of x; x ij : ith data point s jth attribute Let {(x 1, y 1 ), (x, y )(x N, y N )} be the given data, x i D and y i Y Here D

More information

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 Chapter 11 Taylor Series Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 First-Order Approximation We want to approximate function f by some simple function. Best possible approximation

More information

Factors of Polynomials Factoring For Experts

Factors of Polynomials Factoring For Experts Factors of Polynomials SUGGESTED LEARNING STRATEGIES: Shared Reading, Activating Prior Knowledge, Discussion Group, Note-taking When you factor a polynomial, you rewrite the original polynomial as a product

More information

APPENDIX : PARTIAL FRACTIONS

APPENDIX : PARTIAL FRACTIONS APPENDIX : PARTIAL FRACTIONS Appendix : Partial Fractions Given the expression x 2 and asked to find its integral, x + you can use work from Section. to give x 2 =ln( x 2) ln( x + )+c x + = ln k x 2 x+

More information

Math Analysis Notes Mrs. Atkinson 1

Math Analysis Notes Mrs. Atkinson 1 Name: Math Analysis Chapter 7 Notes Day 6: Section 7-1 Solving Systems of Equations with Two Variables; Sections 7-1: Solving Systems of Equations with Two Variables Solving Systems of equations with two

More information

FFTs in Graphics and Vision. The Laplace Operator

FFTs in Graphics and Vision. The Laplace Operator FFTs in Graphics and Vision The Laplace Operator 1 Outline Math Stuff Symmetric/Hermitian Matrices Lagrange Multipliers Diagonalizing Symmetric Matrices The Laplacian Operator 2 Linear Operators Definition:

More information

The Derivative. Appendix B. B.1 The Derivative of f. Mappings from IR to IR

The Derivative. Appendix B. B.1 The Derivative of f. Mappings from IR to IR Appendix B The Derivative B.1 The Derivative of f In this chapter, we give a short summary of the derivative. Specifically, we want to compare/contrast how the derivative appears for functions whose domain

More information

Mathematics (MEI) Advanced Subsidiary GCE Core 1 (4751) June 2010

Mathematics (MEI) Advanced Subsidiary GCE Core 1 (4751) June 2010 Link to past paper on OCR website: www.ocr.org.uk The above link takes you to OCR s website. From there you click QUALIFICATIONS, QUALIFICATIONS BY TYPE, AS/A LEVEL GCE, MATHEMATICS (MEI), VIEW ALL DOCUMENTS,

More information

Lecture V. Numerical Optimization

Lecture V. Numerical Optimization Lecture V Numerical Optimization Gianluca Violante New York University Quantitative Macroeconomics G. Violante, Numerical Optimization p. 1 /19 Isomorphism I We describe minimization problems: to maximize

More information

MULTI VARIABLE OPTIMIZATION

MULTI VARIABLE OPTIMIZATION MULI VARIABLE OPIMIZAION Min f(x 1, x 2, x 3,----- x n ) UNIDIRECIONAL SEARCH - CONSIDER A DIRECION S r r x( α ) = x +α s - REDUCE O Min f (α) - SOLVE AS A SINGLE VARIABLE PROBLEM Min Point s r Uni directional

More information

12.1 Taylor Polynomials In Two-Dimensions: Nonlinear Approximations of F(x,y) Polynomials in Two-Variables. Taylor Polynomials and Approximations.

12.1 Taylor Polynomials In Two-Dimensions: Nonlinear Approximations of F(x,y) Polynomials in Two-Variables. Taylor Polynomials and Approximations. Week 12 Lecture Overheads Math2080 Week 12 Page 1 12.1 Taylor Polynomials In Two-Dimensions: Nonlinear Approximations of F(x,y) Polynomials in Two-Variables. Taylor Polynomials and Approximations. 12.2

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

Lecture 8. Strong Duality Results. September 22, 2008

Lecture 8. Strong Duality Results. September 22, 2008 Strong Duality Results September 22, 2008 Outline Lecture 8 Slater Condition and its Variations Convex Objective with Linear Inequality Constraints Quadratic Objective over Quadratic Constraints Representation

More information

8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0

8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0 8.7 Taylor s Inequality Math 00 Section 005 Calculus II Name: ANSWER KEY Taylor s Inequality: If f (n+) is continuous and f (n+) < M between the center a and some point x, then f(x) T n (x) M x a n+ (n

More information

TEST CODE: MMA (Objective type) 2015 SYLLABUS

TEST CODE: MMA (Objective type) 2015 SYLLABUS TEST CODE: MMA (Objective type) 2015 SYLLABUS Analytical Reasoning Algebra Arithmetic, geometric and harmonic progression. Continued fractions. Elementary combinatorics: Permutations and combinations,

More information

Outline. Scientific Computing: An Introductory Survey. Optimization. Optimization Problems. Examples: Optimization Problems

Outline. Scientific Computing: An Introductory Survey. Optimization. Optimization Problems. Examples: Optimization Problems Outline Scientific Computing: An Introductory Survey Chapter 6 Optimization 1 Prof. Michael. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction

More information

Chapter 7: Exponents

Chapter 7: Exponents Chapter : Exponents Algebra Chapter Notes Name: Algebra Homework: Chapter (Homework is listed by date assigned; homework is due the following class period) HW# Date In-Class Homework M / Review of Sections.-.

More information

OCR Maths FP1. Topic Questions from Papers. Complex Numbers. Answers

OCR Maths FP1. Topic Questions from Papers. Complex Numbers. Answers OCR Maths FP1 Topic Questions from Papers Complex Numbers Answers PhysicsAndMathsTutor.com . 1 (i) i Correct real and imaginary parts z* = i 1i Correct conjugate seen or implied Correct real and imaginary

More information

AUGMENTED LAGRANGIAN METHODS FOR CONVEX OPTIMIZATION

AUGMENTED LAGRANGIAN METHODS FOR CONVEX OPTIMIZATION New Horizon of Mathematical Sciences RIKEN ithes - Tohoku AIMR - Tokyo IIS Joint Symposium April 28 2016 AUGMENTED LAGRANGIAN METHODS FOR CONVEX OPTIMIZATION T. Takeuchi Aihara Lab. Laboratories for Mathematics,

More information

Optimization with Scipy (2)

Optimization with Scipy (2) Optimization with Scipy (2) Unconstrained Optimization Cont d & 1D optimization Harry Lee February 5, 2018 CEE 696 Table of contents 1. Unconstrained Optimization 2. 1D Optimization 3. Multi-dimensional

More information

Computational Optimization. Mathematical Programming Fundamentals 1/25 (revised)

Computational Optimization. Mathematical Programming Fundamentals 1/25 (revised) Computational Optimization Mathematical Programming Fundamentals 1/5 (revised) If you don t know where you are going, you probably won t get there. -from some book I read in eight grade If you do get there,

More information

Math 46 Final Exam Review Packet

Math 46 Final Exam Review Packet Math 46 Final Exam Review Packet Question 1. Perform the indicated operation. Simplify if possible. 7 x x 2 2x + 3 2 x Question 2. The sum of a number and its square is 72. Find the number. Question 3.

More information

MTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1.

MTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1. MTH4101 CALCULUS II REVISION NOTES 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) 1.1 Introduction Types of numbers (natural, integers, rationals, reals) The need to solve quadratic equations:

More information

CAAM 335 Matrix Analysis Planar Trusses

CAAM 335 Matrix Analysis Planar Trusses CAAM 5 Matrix Analysis Planar Trusses September 1, 010 1 The Equations for the Truss We consider trusses with m bars and n nodes. Each node can be displaced in horizontal and vertical direction. If the

More information

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS SECTION C: CONTINUOUS OPTIMISATION LECTURE : THE METHOD OF LAGRANGE MULTIPLIERS HONOUR SCHOOL OF MATHEMATICS OXFORD UNIVERSITY HILARY TERM 005 DR RAPHAEL HAUSER. Examples. In this lecture we will take

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Partial Fraction Decomposition

Partial Fraction Decomposition Partial Fraction Decomposition As algebra students we have learned how to add and subtract fractions such as the one show below, but we probably have not been taught how to break the answer back apart

More information

Stat 206: Linear algebra

Stat 206: Linear algebra Stat 206: Linear algebra James Johndrow (adapted from Iain Johnstone s notes) 2016-11-02 Vectors We have already been working with vectors, but let s review a few more concepts. The inner product of two

More information

Chapter 4: Radicals and Complex Numbers

Chapter 4: Radicals and Complex Numbers Chapter : Radicals and Complex Numbers Section.1: A Review of the Properties of Exponents #1-: Simplify the expression. 1) x x ) z z ) a a ) b b ) 6) 7) x x x 8) y y y 9) x x y 10) y 8 b 11) b 7 y 1) y

More information

Lecture 18: Optimization Programming

Lecture 18: Optimization Programming Fall, 2016 Outline Unconstrained Optimization 1 Unconstrained Optimization 2 Equality-constrained Optimization Inequality-constrained Optimization Mixture-constrained Optimization 3 Quadratic Programming

More information

2.3 Linear Programming

2.3 Linear Programming 2.3 Linear Programming Linear Programming (LP) is the term used to define a wide range of optimization problems in which the objective function is linear in the unknown variables and the constraints are

More information

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Taylor s Theorem Can often approximate a function by a polynomial The error in the approximation

More information

Chapter 2: Unconstrained Extrema

Chapter 2: Unconstrained Extrema Chapter 2: Unconstrained Extrema Math 368 c Copyright 2012, 2013 R Clark Robinson May 22, 2013 Chapter 2: Unconstrained Extrema 1 Types of Sets Definition For p R n and r > 0, the open ball about p of

More information

1 Numerical optimization

1 Numerical optimization Contents 1 Numerical optimization 5 1.1 Optimization of single-variable functions............ 5 1.1.1 Golden Section Search................... 6 1.1. Fibonacci Search...................... 8 1. Algorithms

More information

LECTURE 22: SWARM INTELLIGENCE 3 / CLASSICAL OPTIMIZATION

LECTURE 22: SWARM INTELLIGENCE 3 / CLASSICAL OPTIMIZATION 15-382 COLLECTIVE INTELLIGENCE - S19 LECTURE 22: SWARM INTELLIGENCE 3 / CLASSICAL OPTIMIZATION TEACHER: GIANNI A. DI CARO WHAT IF WE HAVE ONE SINGLE AGENT PSO leverages the presence of a swarm: the outcome

More information

COS 424: Interacting with Data

COS 424: Interacting with Data COS 424: Interacting with Data Lecturer: Rob Schapire Lecture #14 Scribe: Zia Khan April 3, 2007 Recall from previous lecture that in regression we are trying to predict a real value given our data. Specically,

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

Chapter 2 Examples of Optimization of Discrete Parameter Systems

Chapter 2 Examples of Optimization of Discrete Parameter Systems Chapter Examples of Optimization of Discrete Parameter Systems The following chapter gives some examples of the general optimization problem (SO) introduced in the previous chapter. They all concern the

More information

FALL 2018 MATH 4211/6211 Optimization Homework 1

FALL 2018 MATH 4211/6211 Optimization Homework 1 FALL 2018 MATH 4211/6211 Optimization Homework 1 This homework assignment is open to textbook, reference books, slides, and online resources, excluding any direct solution to the problem (such as solution

More information

Flexural-Torsional Buckling of General Cold-Formed Steel Columns with Unequal Unbraced Lengths

Flexural-Torsional Buckling of General Cold-Formed Steel Columns with Unequal Unbraced Lengths Proceedings of the Annual Stability Conference Structural Stability Research Council San Antonio, Texas, March 21-24, 2017 Flexural-Torsional Buckling of General Cold-Formed Steel Columns with Unequal

More information

Reading Mathematical Expressions & Arithmetic Operations Expression Reads Note

Reading Mathematical Expressions & Arithmetic Operations Expression Reads Note Math 001 - Term 171 Reading Mathematical Expressions & Arithmetic Operations Expression Reads Note x A x belongs to A,x is in A Between an element and a set. A B A is a subset of B Between two sets. φ

More information

Core Mathematics 3 Algebra

Core Mathematics 3 Algebra http://kumarmathsweeblycom/ Core Mathematics 3 Algebra Edited by K V Kumaran Core Maths 3 Algebra Page Algebra fractions C3 The specifications suggest that you should be able to do the following: Simplify

More information

MATH 31B: MIDTERM 2 REVIEW. sin 2 x = 1 cos(2x) dx = x 2 sin(2x) 4. + C = x 2. dx = x sin(2x) + C = x sin x cos x

MATH 31B: MIDTERM 2 REVIEW. sin 2 x = 1 cos(2x) dx = x 2 sin(2x) 4. + C = x 2. dx = x sin(2x) + C = x sin x cos x MATH 3B: MIDTERM REVIEW JOE HUGHES. Evaluate sin x and cos x. Solution: Recall the identities cos x = + cos(x) Using these formulas gives cos(x) sin x =. Trigonometric Integrals = x sin(x) sin x = cos(x)

More information

Second Order Optimality Conditions for Constrained Nonlinear Programming

Second Order Optimality Conditions for Constrained Nonlinear Programming Second Order Optimality Conditions for Constrained Nonlinear Programming Lecture 10, Continuous Optimisation Oxford University Computing Laboratory, HT 2006 Notes by Dr Raphael Hauser (hauser@comlab.ox.ac.uk)

More information

9A. Iteration with range. Topics: Using for with range Summation Computing Min s Functions and for-loops A Graphics Applications

9A. Iteration with range. Topics: Using for with range Summation Computing Min s Functions and for-loops A Graphics Applications 9A. Iteration with range Topics: Using for with range Summation Computing Min s Functions and for-loops A Graphics Applications Iterating Through a String s = abcd for c in s: print c Output: a b c d In

More information

1. Method 1: bisection. The bisection methods starts from two points a 0 and b 0 such that

1. Method 1: bisection. The bisection methods starts from two points a 0 and b 0 such that Chapter 4 Nonlinear equations 4.1 Root finding Consider the problem of solving any nonlinear relation g(x) = h(x) in the real variable x. We rephrase this problem as one of finding the zero (root) of a

More information

Numerical optimization

Numerical optimization Numerical optimization Lecture 4 Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Numerical geometry of non-rigid shapes Stanford University, Winter 2009 2 Longest Slowest Shortest Minimal Maximal

More information

8.4 Partial Fractions

8.4 Partial Fractions 8.4 1 8.4 Partial Fractions Consider the following integral. (1) 13 2x x 2 x 2 dx How might we evaluate this? Suppose that, by some good luck, we knew that (2) 13 2x x 2 x 2 = 3 x 2 5 x+1 We could then

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information