Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming Problems

Similar documents
Optimization II: Unconstrained Multivariable

Convex Optimization CMU-10725

5 Quasi-Newton Methods

2. Quasi-Newton methods

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method

Methods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent

Optimization II: Unconstrained Multivariable

Statistics 580 Optimization Methods

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23

Chapter 4. Unconstrained optimization

Nonlinear Optimization: What s important?

Quasi-Newton Methods

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno February 6, / 25 (BFG. Limited memory BFGS (L-BFGS)

Nonlinear Programming

Lecture 7 Unconstrained nonlinear programming

Lecture V. Numerical Optimization

1 Numerical optimization

MATH 4211/6211 Optimization Quasi-Newton Method

Unconstrained optimization

Multivariate Newton Minimanization

Programming, numerics and optimization

1 Numerical optimization

Introduction to unconstrained optimization - direct search methods

Gradient-Based Optimization

ENSIEEHT-IRIT, 2, rue Camichel, Toulouse (France) LMS SAMTECH, A Siemens Business,15-16, Lower Park Row, BS1 5BN Bristol (UK)

Quasi-Newton Methods. Javier Peña Convex Optimization /36-725

Convex Optimization. Problem set 2. Due Monday April 26th

Algorithms for Constrained Optimization

1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by:

Optimization Methods

Higher-Order Methods

Quasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization

Optimization and Root Finding. Kurt Hornik

STAT Advanced Bayesian Inference

ECS550NFB Introduction to Numerical Methods using Matlab Day 2

MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year

Lecture 18: November Review on Primal-dual interior-poit methods

Optimization Tutorial 1. Basic Gradient Descent

SECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS

Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Lecture 14: October 17

Optimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09

NonlinearOptimization

P1: JYD /... CB495-08Drv CB495/Train KEY BOARDED March 24, :7 Char Count= 0 Part II Estimation 183

LBFGS. John Langford, Large Scale Machine Learning Class, February 5. (post presentation version)

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods

Unconstrained Multivariate Optimization

Data Mining (Mineria de Dades)

Quasi-Newton Methods

Chapter 2. Optimization. Gradients, convexity, and ALS

Gradient Descent. Dr. Xiaowei Huang

Improving the Convergence of Back-Propogation Learning with Second Order Methods

230 L. HEI if ρ k is satisfactory enough, and to reduce it by a constant fraction (say, ahalf): k+1 = fi 2 k (0 <fi 2 < 1); (1.7) in the case ρ k is n

(One Dimension) Problem: for a function f(x), find x 0 such that f(x 0 ) = 0. f(x)

Optimization Methods for Machine Learning

Numerical solutions of nonlinear systems of equations

OPER 627: Nonlinear Optimization Lecture 14: Mid-term Review

Introduction to gradient descent

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2

Constrained optimization: direct methods (cont.)

Part 4: IIR Filters Optimization Approach. Tutorial ISCAS 2007

Search Directions for Unconstrained Optimization

Optimization. Totally not complete this is...don't use it yet...

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

MATH 4211/6211 Optimization Basics of Optimization Problems

Line Search Methods for Unconstrained Optimisation

17 Solution of Nonlinear Systems

ORIE 6326: Convex Optimization. Quasi-Newton Methods

Introduction to Nonlinear Optimization Paul J. Atzberger

Review of Classical Optimization

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)

Lecture Notes: Geometric Considerations in Unconstrained Optimization

Motivation: We have already seen an example of a system of nonlinear equations when we studied Gaussian integration (p.8 of integration notes)

Quasi-Newton methods for minimization

Extra-Updates Criterion for the Limited Memory BFGS Algorithm for Large Scale Nonlinear Optimization 1

Improved Damped Quasi-Newton Methods for Unconstrained Optimization

CE 191: Civil and Environmental Engineering Systems Analysis. LEC 05 : Optimality Conditions

Optimization Methods

ECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor.

Geometry optimization

Static unconstrained optimization

Regression with Numerical Optimization. Logistic

Numerical Analysis of Electromagnetic Fields

Chapter 3 Numerical Methods

Matrix Derivatives and Descent Optimization Methods

ON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS

Numerical Optimization: Basic Concepts and Algorithms

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey

ECON 5350 Class Notes Nonlinear Regression Models

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

Topic 8c Multi Variable Optimization

Step lengths in BFGS method for monotone gradients

8 Numerical methods for unconstrained problems

An Algorithm for Unconstrained Quadratically Penalized Convex Optimization (post conference version)

Global Convergence of Perry-Shanno Memoryless Quasi-Newton-type Method. 1 Introduction

Transcription:

International Journal of Scientific and Research Publications, Volume 3, Issue 10, October 013 1 ISSN 50-3153 Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming Problems Neha Varma*, Dr. Ganesh Kumar** * (NET- JRF, Department of Mathematics, B.R.A. Bihar University, Muzaffarpur, Bihar, India) ** (University Professor& Ex. H.O.D. of Mathematics, B.R.A. Bihar University, Muzaffarpur, Bihar, India) Abstract- In this paper we propose to discuss unconstrained multivariable search methods that are used for optimization of nonlinear programming problems. Earlier Dr.William P.Fox and Dr.William H.Richardson [1] have attempted to solve such problems by using MAPPLE. But we have preferred to solve these problems by OR methods as our present area of research is OR. Although several OR methods are known but we have confined our discussions to Gradient search method, Newton s method and Quasi- Newton methods. We propose to conclude the discussion by taking a suitable example.. Key Words- Multivariable,Optimization, Quasi -Newton methods, steepest ascent/descent P I. INTRODUCTION roblems containing more than one variable are called multivariate. The univariate results have multivariate analogues. In the multivariate case, again the necessary and sufficient condition for optimality is given by the system of equations obtained by setting the respective partial derivatives equal to zero. The first and second order partial derivatives are again key to the optimization process, except that a vector of the first derivatives and a matrix of second derivatives are involved. II. Gradient Search Method First of all, we take up the gradient search method also known as steepest ascent/descent method which is the simplest and most fundamental method for unconstrained optimization. When the negative gradient is used as its descent direction, the methd is known as steepest descent method and when the positive gradient is used as its ascent direction, the method is called steepest ascent method. The gradient search method can be summarized in the following steps: 1) Choose an initial starting point x 0. Thereafter at the point x k : ) Calculate (analytically or numerically) the partial derivatives f(x)/ x j, j 1,, n 3) Calculate the maximum/minimum of the new function f(x k ±t k f(x k )) by using one dimensional search procedure (or calculus method) to find tt k that maximizes/minimizes f(x k ±t k minimization). f(x k )) over t k 0 (for maximization) or t k 0 (for 4) Reset x k+1 x k +t k f(x k ) (for maximization), x k+1 x k -t k f(x k ) (for minimization). Then go to the stopping rule. Stopping rule: At the maximum/ minimum, the value of the elements of the gradient vector will be each equal to zero. So we evaluate f(x k ) at xx k. If f/ x j ϵ for all j 1,,3.n we must stop with the current x k as the desired approximation of an optimal solution. If not, repeat above steps and set k k+1. III. Newton s method Newton s method for multivariable optimization is analogues to Newton s single variable algorithm for obtaining the roots and Newton Raphson method for finding the roots of first derivative, given a x 0, iterates x k+1 x k - f (x k )/f (x k ) until Ix k+1 - x k I ϵ

International Journal of Scientific and Research Publications, Volume 3, Issue 10, October 013 ISSN 50-3153 The algorithm is expanded to include partial derivatives w.r.t. each variable s dimension. The algorithm is developed by locating the stationary point of equation f(x) f(x k )+ f(x k ) T (x-x k )+ ½(x-x k ) T H(x-x k ) where H is the Hessian matrix and ( x-x k ) T is the row vector, which is the transpose of the column vector of the difference between the vector of independent variables x and the point x k used for Taylor s series expansionby setting the first partial derivatives with respect to x 1, x,.x n equal to zero i.e. f(x k )+ H(x- x k ) 0 Then solving for x, the optimum of the quadratic approximation, Newton s method algorithm is obtained as X k+1 x k - H -1 f(x k ) The Newton s method consists of the following steps: 1) Choose an initial point x 0. ) Calculate (analytically or numerically) the partial derivatives f(x)/ x j, j 1,,3,n 3) Calculate the Hessian matrix H k, the matrix of second partial derivatives at the point x k. 4)Find the inverse of H K i.e. H K -1. 5) Set x k+1 x k - H -1 f(x k ) 6) If <ϵthen stop otherwise repeat the above steps. IV. Quasi -Newton methods -1 We have noticed that Newton s method x k+1 x k + H K f(x k ) (for maximization) is successful because it uses the Hessian which offers the useful curvature information. However for various practical problems, the computing efforts of the Hessian matrices are very expensive or the evaluation of Hessian is very difficult. Sometimes the Hessian is not available analytically. To overcome these disadvantages Quasi-Newton methods were developed. These methods use the functional values and the gradients of the objective function and also at the same time maintains a fast rate of convergence. There are several updates of Quasi-Newton method. Here we confine our discussion only to DFP update and BFGS update which are both rank two updates. DFP update DFP update is a rank- two update and has become the best known of the Quasi-Newton algorithms. The DFP update formula was originally proposed by Davidon [6] and developed later by Fletcher and Powell [7]. Hence it is called DFP update. This formula preserves the positive definiteness in case of minimization and negative definiteness in case of maximization and also symmetry of matrices H k. DFP method is superlinearly convergent. But for a strictly convex function, under exact line search, DFP method is globally convergent. The DFP algorithm has the following form of equation for maximizing f(x): x k+1 x k + t k H K f(x k ) where H k H k-1 + A k + B k and the matrices A k and B k are given by A k {(x k - x k-1 ) (x k - x k-1 ) T }/{(x k - x k-1 ) T ( f(x k )- f(x k-1 )} B k {-H k-1 ( f(x k )- f(x k-1 ) )( f(x k )- f(x k-1 )) T H T k-1 }/{ ( f(x k )- f(x k-1 )) T H k-1 ( f(x k )- f(x k-1 ))} The algorithm begins with a search along the gradient line from the starting point x 0 as given by the following equation x 1 x 0 + t 0 H 0 f(x 0 ) where H 0 I is the unit matrix. BFGS update The other famous Quasi-Newton update BFGS update overcomes all the drawbacks ofoter methods discussed previously and performs better than DFP update. The BFGS update formula developed simultaneously by Broyden[8], Fletcher[9], Goldfarb[10], Shanno[11] in the year 1970 and hence known as BFGS formula. The BFGS algorithm for maximizing f(x) is given by x k+1 x k + t k H k f(x k ) where H k H k-1 - A k + B k The matrices A k and B k are given by A k {H k-1 ϒ k δ k T + δ k ϒ k T H k-1 }/ {δ k T ϒ k } B k {1+ ϒ k T H k-1 ϒ k / δ k T ϒ k }{ δ k δ k T / δ k T ϒ k } where δ k x k+1 x k and ϒ k f(x k+1 ) - f(x k ) Here also x 1 is calculated as in DFP update formula.

International Journal of Scientific and Research Publications, Volume 3, Issue 10, October 013 3 ISSN 50-3153 V. Illustration: Maximize f(x) x 1 x + x x 1 x using Gradient search starting at point x 0 (0, 0). Here f(x) x 1 x + x x 1 x f(x 1, x ) T (-x 1 +x, x 1-4x +) f(x 0 ) T f(0,0) T (0,) x 1 x 0 + t 0 f(x 0 ) + t 0 x 1 T (0,t 0 ) Now f(t) 4t 0 8t 0 For optimum value, f (t) 0 t 0 1/4 Also f (t) -16<0 f(t) is maximum at t 0 1/4 x 1 T (0, 1/) Proceeding as above we get a sequence of iterates as x (1/,1/), x 3 (1/,3/4), x 4 (3/4,3/4), x 5 (3/4,7/8), x 6 (7/8,7/8), x 7 (7/8,15/16) and so on. We observe that these sequence of trials are converging to x* (1,1) which is the optimal solution as verified by the analytical method. By Newton s method We have f(x) x 1 x +x x 1 -x Now H 0 H 0 is negative definite. Also H 0 (8-4) 4 f(x 1,x ) T (-x 1 +x, x 1-4x + ) f(0,0) T (0,) H -1 0 T -1 x 1 x 0 H 0 f(x 0 ) (1,1) is the required point of maxima. Clearly the optimum of this quadratic function is obtained in one step as stated in the method. By DFG update formula f(x) x 1 x + x x 1 -x f(x 1,x ) T (-x 1 + x, x 1-4x +) f(x 0 ) f(0,0) T (0,) Now, x 1 x 0 + t 0 H 0 (I) f(x 0 ) + t 0 x 1 T (0, t 0 ) Now f(t) 4t 0 8t 0 For optimum value, f (t) 0 t 0 1/4 Also f (t) -16 < 0 f(t) is maximum at t 0 1/4 x 1 T (0, 1/) f(x 1 ) T (1,0) Now x x 1 + t 1 H 1 f(x 1 ) where H 1 H 0 + A 1 + B 1

International Journal of Scientific and Research Publications, Volume 3, Issue 10, October 013 4 ISSN 50-3153 Also A 1 B 1 - - - H 1 x T (4t 1 /5,1/ + t 1 /5 ) Obtaining t 1 by an exact line search, we get x T (1, 1) which is the required point of maxima. We observe that for a quadratic function with two independent variables,the method converges to the optimum after two iterations. By BFGS update formula Here f(x) x 1 x + x x x f(x 1,x ) T ( -x 1 + x, x 1-4x +) f(x 0 ) T f(0,0) (0, ) Now x 1 T x 0 + t o H(I) f(x 0 ) (0, t 0 ) Let f(t) 4t 0 8t 0 For optimum value, f (t)0 t 0 1/4 Also f (t) -16< 0 f(t) is maximum at t 0 1/4 x 1 T (0, 1/) f(x 1 ) (1,0) Now x x 1 + t 1 H 1 f(x 1 ) where H 1 H 0 A 1 + B 1 A 1 B 1 {1+ }X H 1 - +

International Journal of Scientific and Research Publications, Volume 3, Issue 10, October 013 5 ISSN 50-3153 Now x + t 1 x T (t 1, 1/+ 1/ t 1 ) Obtaining t 1 by an exact line search, we get x T (1, 1) which is the required point of maxima. VI. Conclusion Surprisingly not much attention has been given to maximization of multivariable nonlinear programming problems by the scholars. So we have chosen to study maximization problems and to our pleasant surprise, the results obtained are compatible with theoretical observations. In gradient search methods, the rate of convergence is slow and the result obtained is approximate to the optimal value. But in Newton s method and Quasi-Newton methods, the rate of convergence is faster and the results are accurate. Acknowledgement To all research scholars of this field and to Dr.William P.Fox and Dr.William H.Richardson in particular, whose works motivated us in preparing this research paper. References [1] Dr.William P. Fox and Dr.William H.Richardson, Department of Mathematics, Francis Marion University Florence,SC 9501; Comparison of NLP search methods in Nonlinear Optimization using MAPPLE. [] Frederick S. Hillier, Gerald J.Lieberman; Introduction to Operations Research, Tata Mc Graw-Hill Publishing company limited. [3] Wenyu Sun and YA-Xiang Yuan; Optimization theory and methods, Nonlinear Programming, Springer. [4] M.Avriel; Nonlinear Programming: Analysis and methods, Prentice-Hall, Inc, 1976. [5] R.Fletcher, John Willey and sons, Chichester; Practical methods of Optimization, Vol.1, Unconstrained Optimization, 1981. [6] W.C.Davidon; Variable metric methods for minimization, Argonne National Labs Report, ANL-5990, (1959). [7] R.Fletcher and M.J.D.Powell; A rapid convergent descent method for minimization, Computer Journal 6(1963) 163-168. [8] The convergence of a class double-rank minimization algorithms, Journal of the Institute of Mathematics and its Applications 6(1970) 76-90. [9] R.Fletcher ; A new approach to variable metric algorithms, Computer J.13(1970) 317-3. [10] D.Goldfarb.; A family of variable metric methods derived by variation mean, Mathematics of Computation 3(1970) 3-6. [11] D.F.Shanno; Conditioning of Quasi-Newton methods for function minimization, Math. Comput.4(1970) 647-656. AUTHORS Neha Varma, NET- JRF, Department of Mathematics, B.R.A. Bihar University, Muzaffarpur, Bihar, India raginineha.v@gmail.com Dr. Ganesh Kumar, University Professor& Ex. H.O.D. of Mathematics, B.R.A. Bihar University, Muzaffarpur, Bihar, India. Correspondence Author. Neha Varma, NET- JRF, Department of Mathematics, B.R.A. Bihar University, Muzaffarpur, Bihar, India raginineha.v@gmail.com, sp.varmashailesh@gmail.com, 09835815550