# Exploring the energy landscape

Save this PDF as:

Size: px
Start display at page:

## Transcription

1 Exploring the energy landscape ChE210D Today's lecture: what are general features of the potential energy surface and how can we locate and characterize minima on it Derivatives of the potential energy function Forces For any atomic configuration, we can compute the force acting on each atom from the negative energy derivative with respect to the particle coordinates: In shorthand notation, Alternatively,,,, That is, the force stems from the gradient of the potential energy. We can simplify this expression if our potential energy function is built from pairwise interactions, We have that: + + Consider computing the force on atom. There are 1 terms in the pairwise summation that include the coordinates of atom ". Thus the energy derivative with respect to is:, # M. S. Shell /14 last modified 4/12/2012

2 # Using the definition of, 2 2 Hence,, Doing the same for the and coordinates, where # # For example, if our pair potential is the so-called soft-sphere interaction, where + is a positive integer, then Simplifying, %& ' ( )*, +%' * )*)-. # + )* /%& ' ( 0 # Notice that the force acting on atom " due to atom 1 is just the negative of the force acting on atom 1 due to atom ", since 2 M. S. Shell /14 last modified 4/12/2012

3 In simulation, we typically maintain an (N,3) array of forces. A typical way to implement pairwise force and potential energy calculation in computer code is the following: set the potential energy variable and elements of the force array equal to zero loop over atom i 1 to N 1 loop over atom j i+1 to N calculate the pair energy of atoms i and j add the energy to the potential energy variable calculate the force on atom i due to j add the force to the array elements for atom i subtract the force from the array elements for atom j Second derivatives and the Hessian The curvature of the potential energy function surrounding a given atomic configuration is given by second derivatives. There are 3 3 different second derivatives, attained by cross-permuting different particle coordinates. We typically write the second derivatives as a matrix called the Hessian: 8 > < The Hessian is important in distinguishing potential energy minima from saddle points, as discussed in greater detail below. It can be a memory-intensive matrix to compute in molecular simulation, as the number of elements is 9. The energy landscape Basic definition and properties The potential energy function defines all of the thermodynamic and kinetic properties of a classical atomic system. We can think of this function as being plotted in a high-dimensional space where there are 3 axes, one for each coordinate of every atom. An additional axis measures the potential energy, and we call the projection of the potential energy function in 3 +1 dimensional space as the potential energy landscape (PEL). Though we cannot visualize the PEL directly, the following schematic shows some features of it: M. S. Shell /14 last modified 4/12/2012

4 saddle point basin local minima particle coordinates, global minimum Stationary points correspond to the condition where the slope of the PEL is This is shorthand for the following simultaneous conditions: "1,2,, Notice that these equations imply that the net force acting on each particle is zero: 0 Therefore, the stationary points correspond to mechanically stable configurations of the atoms in the system. There are two basic kinds of stationary points: For minima, the curvature surrounding the stationary point is positive. That is, any movement away from the point increases the potential energy. For saddles, one can move in one or more directions around a stationary point and experience a decrease in potential energy. Scaling of the number of minima and saddle points PELs contain numerous stationary points. It can be shown [Stillinger, Phys Rev E (1999); Shell et al, Phys Rev Lett (2004)] that both the number of minima and saddle points grow exponentially with the number of atoms: number of minima ~exp ' M. S. Shell /14 last modified 4/12/2012

5 number of saddle points ~ exp T where ' and T are system-specific constants but independent of the number of atoms. The implication of these scaling laws is that it is extraordinarily difficult to search exhaustively for all of the minima for even very small systems (~100 atoms), since their number is so great. By extension, it is extremely challenging to locate the global potential energy minimum, that is, the set of particle coordinates that gives the lowest potential energy achievable. The following figure shows a disconnectivity graph for a system of just 13 Lennard-Jones atoms in infinite volume, taken from [Doye, Miller, and Wales, J. Chem. Phys. 111, 8417 (1999)]. The horizontal axis marks the potential energy and the lines show the connectivity of different minima. To move from one minimum to another, one must travel the path formed when the lines extending from two minima connect to a common energetic point. This small system has 1467 distinct potential energy minima. These results were generated using an advanced and computationally-intensive minima-finding algorithm. Relevance of finding minima Why should we care to find PEL minima? M. S. Shell /14 last modified 4/12/2012

6 Mechanically stable molecular structures can be identified. When examining conformational changes within a molecule, minima give the distinct conformational states possible. For changes among many molecules, the minima provide the basic structures associated with physical events in the system (e.g., binding). Both can often be compared with experiment. Differences in the energies of different minimized structures provide a first-order perspective on the relative populations of these (neglecting entropies). Again, this can be compared to experiment and often is an important force field diagnostic. Initial structures in molecular simulations often need to be relaxed away from highenergy states. Minimization can take an initial structure and remove atomic overlaps or distorted bond and angle degrees of freedom. Locating local energy minima Problem formulation If we are anywhere within the basin of a minimum, we can define a steepest descent protocol that takes us to the minimum: U limu ) ( ) Here, U is a fictitious time-like variable. The solution to this first order differential equation in the limit that U is the set of coordinates at a minimum min. Here, we essentially follow the forces down the potential energy surface until we reach its bottom the velocities vectors of each atom follow the negative forces. In a two-dimensional contour plot of energies (as a function of coordinates), this would look like: M. S. Shell /14 last modified 4/12/2012

7 The steepest descent protocol is equivalent to an instantaneous quench of a system to absolute zero: we continuously remove energy from the system until we cannot remove any more. Line searches It is very inefficient to implement the above equation because one must take U to infinity, and because the closer we get to the minimum, the slower we are to converge due to the fact that the forces grow increasingly small. A number of much more efficient methods are available. These typically use a so-called line search as a part of the overall approach. A line search simply says to find a minimum along one particular direction in coordinate space, X. In two dimensions, X The procedure for a line search is broken into two parts. 1. Start with an initial set of coordinates Y and a search direction X, chosen to be in the downhill direction of the energy surface. 2. Let - Y. Then, find two more coordinates Y +ZX and [ Y +2ZX by taking small steps Z along the search direction. Z should be chosen to be small relative to the scales of energy changes relative to particle coordinates. 3. Is ( [ )>( )? If so, the three coordinates bracket a minimum and we should move to the next part of the algorithm. 4. If not, we need to keep searching along this direction. Let -, [, and [ +ZX. Go back to step 3. The second part of the algorithm consists of finding the minimum once we ve bracketed it. One possibility is to bisect pairs of coordinates: 1. Find a new coordinate ^. If - > [, pick ^ - ( - + ). Otherwise, pick ^ - ( + [ ). M. S. Shell /14 last modified 4/12/2012

8 2. Of the four coordinates ( -,, [,^) pick new values for ( -,, [ ) from these that most closely bracket the minimum. 3. Go back to step 1 until the successive values of the energies found at each new ^ changes by less than some fractional tolerance, typically 1 in ( ) progress along X progress along X An alternative to step 1 here, which converges faster, is to use the three points ( -,, [ ) to fit a parabola and find the coordinates at the minimum from this approximation: - - -[ [ - Δ - ( ) ( - ) Δ -[ ( [ ) ( - ) ^ a - - ba -[ Δ - - Δ -[ b -[ Δ - - Δ -[ Steepest descent with line search The steepest descent approach can be made more efficient with a line search: 1. From the current set of coordinates, compute the forces. Choose the search direction to be equal to X, the negative gradient of the potential energy surface. 2. Perform a line search to find the minimum along X. 3. Compute the new set of forces at the configuration found in point 2 and a new search direction from them. 4. Repeat step 2 until the minimum energy found with each iteration no longer changes by some fractional tolerance, typically 1 in M. S. Shell /14 last modified 4/12/2012

9 Conjugate-gradient method For long, narrow valleys in the PEL, the steepest descent + line search method results in a large number of zigzag-type motions as one approaches the minimum. This is because the approach continuously overcorrects itself when computing the new search direction with each iteration (new line search). Instead, a more efficient approach is to use the conjugate gradient (CG) method. The CG approach differs only in the choice of a new search direction with each iteration in step 3. Like the steepest descent + line search approach, we use the gradient of the energy surface. In addition, however, we use the previous search direction as well. The CG method chooses search directions that are conjugate (perpendicular) to previous gradients found. A new search direction at the "th iteration is found using with X +c X )- c )- )- Here, the dot product signifies the 3N-dimensional vector multiplication:,- +,- +,- +, + +, For +-dimensional quadratic functions, the CG approach converges exactly to the minimum in + iterations (e.g., + line search operations). An alternative equation for c is typically used in molecular modeling, as it has better convergence properties: c ( )- ) )- )- M. S. Shell /14 last modified 4/12/2012

10 Newton and Quasi-Newton methods Location of minima can be enhanced by using second derivative information. Consider the PEL in the vicinity of a minimum. For the sake of simplicity, we will consider a one-dimensional function (). We can Taylor expand () and its derivative to second order: ()( Y )+( Y ) ( Y) () ( Y) + ( Y) ( Y ) 2 +( Y ) ( Y ) Neglecting higher terms, we can solve for the point at which the derivative is zero: Y a ( )- Y) ( Y ) ba b For quadratic functions, this relation will find the minimum after only a single application. For other functions, this equation defines an iterative procedure whereby the minimum can be located through successive applications of it (each time the function being approximated as quadratic): e- a ( )- ) ( ) ba b This is the Newton-Raphson iteration scheme for finding stationary points. The above procedure will converge to both minima and saddle points. To converge to minima, one must first move to a point on the PEL where the energy landscape has positive curvature in every direction. For multidimensional functions ( ) the update equation above involves an inverse of the Hessian matrix of second derivatives: e- + )- 5 To implement this method in simulation requires computation of the 3 3 Hessian matrix and its inverse with each iteration step. For even small systems, this task can become computeand memory-intensive. To remedy these issues, several quasi-newton methods are available. The distinctive feature of these approaches is that they generate approximations to the inverse Hessian; upon each iteration, these approximations grow increasingly accurate. These approximations are generated from the force array. The most popular of these in use with molecular modeling methods is M. S. Shell /14 last modified 4/12/2012

11 the Limited Broyden-Fletcher-Goldfarb-Shanno (LBFGS) approach. Implementations of this method can be found at online simulation code databases. In general, Newton and quasi-newton methods are more accurate and efficient than the conjugate gradient approach; however, both are widely used in the literature. Locating global energy minima Methods to find global potential energy minima are actively researched and have important applications in crystal and macromolecular structure prediction. The global minimum problem is much more challenging as it requires a detailed search of the entire PEL and it is difficult if not impossible to know a priori when the global minimum has been found (unless the entire surface has been explored exhaustively). Typically global-minimizers combine local minimization with various other methods to selectively explore low-energy regions of the PEL and with bookkeeping routines to remember which parts have been searched and what low-energy configurations have been found. This search is usually partially stochastic in nature, often leveraging the molecular dynamics and Monte Carlo techniques we will discuss in later lectures. Normal mode analysis For minima Consider the PEL in the vicinity of a minimum. Let s write the coordinates of all the atoms about this minimum be If we consider very small perturbations about these coordinates, Y Y +Δ we can Taylor expand the potential energy to second order. That is, we can perform a harmonic approximation to the PEL about a minimum: ( )( Y )+ Δ - 2 ( Y ) + Δ -Δ - ( Y ) + + Δ ( Y ) Notice that all of the first-order terms in the Taylor expansion are zero by the fact that ( Y ) is a minimum. A shorter way of writing this is in matrix form using the Hessian: M. S. Shell /14 last modified 4/12/2012

12 f > 8 Δ - Δ Δ Δ - 27Δ 7 6Δ < which can be written in more compact form: 8 > Δ - 7 Δ > Δ - 7Δ Δ < < Y gf 5g where Y ( Y ) and gδ. Keep in mind that the Hessian is evaluated at Y. We can use a trick of mathematics to simplify the analysis, by diagonalizing the Hessian matrix 5. Notice that 5 is a symmetric matrix, since the order in which the second derivatives are taken doesn t matter. That property means that we can write where i is a diagonal matrix of eigenvalues: 5h f ih k k ij 0 l 0 0 k and h is a matrix whose rows are eigenvectors of 5. Then, Y gf h f ihg Y (hg)f i(hg) Y mf im In the last line, we defined a new set of 3 coordinates, instead of (g), that consists of linear combinations of the atomic positions: mhg h h Y M. S. Shell /14 last modified 4/12/2012

13 Physically, the new coordinates m correspond to projecting the atomic positions on a new set of axes given by the eigenvectors of the Hessian. Using the above Taylor expansion, the fact that i is a diagonal matrix means that the energy is a simple a sum of 3 independent harmonic oscillators along the new directions m: [ Y k U The physical interpretation of this result is the following. Starting from the minimum and at low temperatures, the system oscillates about Y. The eigenvectors of 5 give the normal modes along which the system fluctuates, that is, the directions and collective motions of all of the atoms that move in harmonic fashion for each oscillator. There are 3 of these modes corresponding to the 3 eigenvectors. The eigenvalues give the corresponding force constants and can be related to the frequencies of fluctuations. If all atomic masses are the same, o, then n- p qk o In the case that the atomic masses differ, a slightly more complicated equation is used [Case, Curr. Op. Struct. Biol. 4, 285 (1994)]. Regarding the eigenvalues, All will be positive or zero, since the energy must increase as one moves away from the minimum. Thus, the frequencies are all real. If the system also has rotational symmetry (i.e., atoms can be collectively rotated about the x, y, and z axes), then an additional three eigenvalues will be zero. The spectrum of different normal mode frequencies informs one of motions on different timescales: If the system has translational symmetry (i.e., atoms can collectively be moved in the x, y, and z directions without a change in energy), then three of the eigenvalues will be zero. High-frequency modes signify very fast degrees of freedoms. The corresponding eigenmodes typically correspond to bond stretching motions, if present. Highfrequency modes tend to be localized, i.e., involving only a few atoms. M. S. Shell /14 last modified 4/12/2012

### Unconstrained optimization

Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout

### Lecture Notes: Geometric Considerations in Unconstrained Optimization

Lecture Notes: Geometric Considerations in Unconstrained Optimization James T. Allison February 15, 2006 The primary objectives of this lecture on unconstrained optimization are to: Establish connections

### Physics 403. Segev BenZvi. Numerical Methods, Maximum Likelihood, and Least Squares. Department of Physics and Astronomy University of Rochester

Physics 403 Numerical Methods, Maximum Likelihood, and Least Squares Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Quadratic Approximation

### CPSC 540: Machine Learning

CPSC 540: Machine Learning First-Order Methods, L1-Regularization, Coordinate Descent Winter 2016 Some images from this lecture are taken from Google Image Search. Admin Room: We ll count final numbers

### Optimization Methods

Optimization Methods Decision making Examples: determining which ingredients and in what quantities to add to a mixture being made so that it will meet specifications on its composition allocating available

### Introduction to Geometry Optimization. Computational Chemistry lab 2009

Introduction to Geometry Optimization Computational Chemistry lab 9 Determination of the molecule configuration H H Diatomic molecule determine the interatomic distance H O H Triatomic molecule determine

### Multivariate Newton Minimanization

Multivariate Newton Minimanization Optymalizacja syntezy biosurfaktantu Rhamnolipid Rhamnolipids are naturally occuring glycolipid produced commercially by the Pseudomonas aeruginosa species of bacteria.

### Example questions for Molecular modelling (Level 4) Dr. Adrian Mulholland

Example questions for Molecular modelling (Level 4) Dr. Adrian Mulholland 1) Question. Two methods which are widely used for the optimization of molecular geometies are the Steepest descents and Newton-Raphson

### , b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are

Quadratic forms We consider the quadratic function f : R 2 R defined by f(x) = 2 xt Ax b T x with x = (x, x 2 ) T, () where A R 2 2 is symmetric and b R 2. We will see that, depending on the eigenvalues

### Phonons and lattice dynamics

Chapter Phonons and lattice dynamics. Vibration modes of a cluster Consider a cluster or a molecule formed of an assembly of atoms bound due to a specific potential. First, the structure must be relaxed

### EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Multidimensional Unconstrained Optimization Suppose we have a function f() of more than one

### 22 Path Optimisation Methods

22 Path Optimisation Methods 204 22 Path Optimisation Methods Many interesting chemical and physical processes involve transitions from one state to another. Typical examples are migration paths for defects

### CSC321 Lecture 8: Optimization

CSC321 Lecture 8: Optimization Roger Grosse Roger Grosse CSC321 Lecture 8: Optimization 1 / 26 Overview We ve talked a lot about how to compute gradients. What do we actually do with them? Today s lecture:

### Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

### Vasil Khalidov & Miles Hansard. C.M. Bishop s PRML: Chapter 5; Neural Networks

C.M. Bishop s PRML: Chapter 5; Neural Networks Introduction The aim is, as before, to find useful decompositions of the target variable; t(x) = y(x, w) + ɛ(x) (3.7) t(x n ) and x n are the observations,

### CSC321 Lecture 7: Optimization

CSC321 Lecture 7: Optimization Roger Grosse Roger Grosse CSC321 Lecture 7: Optimization 1 / 25 Overview We ve talked a lot about how to compute gradients. What do we actually do with them? Today s lecture:

### Today. Introduction to optimization Definition and motivation 1-dimensional methods. Multi-dimensional methods. General strategies, value-only methods

Optimization Last time Root inding: deinition, motivation Algorithms: Bisection, alse position, secant, Newton-Raphson Convergence & tradeos Eample applications o Newton s method Root inding in > 1 dimension

### (One Dimension) Problem: for a function f(x), find x 0 such that f(x 0 ) = 0. f(x)

Solving Nonlinear Equations & Optimization One Dimension Problem: or a unction, ind 0 such that 0 = 0. 0 One Root: The Bisection Method This one s guaranteed to converge at least to a singularity, i not

### NonlinearOptimization

1/35 NonlinearOptimization Pavel Kordík Department of Computer Systems Faculty of Information Technology Czech Technical University in Prague Jiří Kašpar, Pavel Tvrdík, 2011 Unconstrained nonlinear optimization,

### Project: Vibration of Diatomic Molecules

Project: Vibration of Diatomic Molecules Objective: Find the vibration energies and the corresponding vibrational wavefunctions of diatomic molecules H 2 and I 2 using the Morse potential. equired Numerical

### Gradient Descent. Dr. Xiaowei Huang

Gradient Descent Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Three machine learning algorithms: decision tree learning k-nn linear regression only optimization objectives are discussed,

### TECHNIQUES TO LOCATE A TRANSITION STATE

32 III. TECHNIQUES TO LOCATE A TRANSITION STATE In addition to the location of minima, the calculation of transition states and reaction pathways is an interesting task for Quantum Chemistry (QC). The

### Deep Learning. Authors: I. Goodfellow, Y. Bengio, A. Courville. Chapter 4: Numerical Computation. Lecture slides edited by C. Yim. C.

Chapter 4: Numerical Computation Deep Learning Authors: I. Goodfellow, Y. Bengio, A. Courville Lecture slides edited by 1 Chapter 4: Numerical Computation 4.1 Overflow and Underflow 4.2 Poor Conditioning

### Mathematical optimization

Optimization Mathematical optimization Determine the best solutions to certain mathematically defined problems that are under constrained determine optimality criteria determine the convergence of the

### Programming, numerics and optimization

Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428

### Neural Network Training

Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification

### Dynamics of Lennard-Jones clusters: A characterization of the activation-relaxation technique

PHYSICAL REVIEW E VOLUME 62, NUMBER 6 DECEMBER 2000 Dynamics of Lennard-Jones clusters: A characterization of the activation-relaxation technique Rachid Malek* and Normand Mousseau Department of Physics

Gradient Descent Sargur srihari@cedar.buffalo.edu 1 Topics Simple Gradient Descent/Ascent Difficulties with Simple Gradient Descent Line Search Brent s Method Conjugate Gradient Descent Weight vectors

### Lecture V. Numerical Optimization

Lecture V Numerical Optimization Gianluca Violante New York University Quantitative Macroeconomics G. Violante, Numerical Optimization p. 1 /19 Isomorphism I We describe minimization problems: to maximize

### Computer simulation can be thought as a virtual experiment. There are

13 Estimating errors 105 13 Estimating errors Computer simulation can be thought as a virtual experiment. There are systematic errors statistical errors. As systematical errors can be very complicated

### Improving the Convergence of Back-Propogation Learning with Second Order Methods

the of Back-Propogation Learning with Second Order Methods Sue Becker and Yann le Cun, Sept 1988 Kasey Bray, October 2017 Table of Contents 1 with Back-Propagation 2 the of BP 3 A Computationally Feasible

### Nonlinear Optimization

Nonlinear Optimization (Com S 477/577 Notes) Yan-Bin Jia Nov 7, 2017 1 Introduction Given a single function f that depends on one or more independent variable, we want to find the values of those variables

### ( ) + ( ) + ( ) = 0.00

Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Spring Semester 2006 Christopher J. Cramer Lecture 32, April 14, 2006 (Some material in this lecture has been adapted from Cramer, C.

### Molecular Simulation II. Classical Mechanical Treatment

Molecular Simulation II Quantum Chemistry Classical Mechanics E = Ψ H Ψ ΨΨ U = E bond +E angle +E torsion +E non-bond Jeffry D. Madura Department of Chemistry & Biochemistry Center for Computational Sciences

### Unconstrained Multivariate Optimization

Unconstrained Multivariate Optimization Multivariate optimization means optimization of a scalar function of a several variables: and has the general form: y = () min ( ) where () is a nonlinear scalar-valued

### Quasi-Newton Methods

Newton s Method Pros and Cons Quasi-Newton Methods MA 348 Kurt Bryan Newton s method has some very nice properties: It s extremely fast, at least once it gets near the minimum, and with the simple modifications

### Nonlinear Optimization for Optimal Control

Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]

### Data Mining (Mineria de Dades)

Data Mining (Mineria de Dades) Lluís A. Belanche belanche@lsi.upc.edu Soft Computing Research Group Dept. de Llenguatges i Sistemes Informàtics (Software department) Universitat Politècnica de Catalunya

### Lecture 35 Minimization and maximization of functions. Powell s method in multidimensions Conjugate gradient method. Annealing methods.

Lecture 35 Minimization and maximization of functions Powell s method in multidimensions Conjugate gradient method. Annealing methods. We know how to minimize functions in one dimension. If we start at

6-1: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction to gradient descent Derivation and intuitions Hessian 6-2: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction Our

### Nonlinear Programming

Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week

### Optimization for neural networks

0 - : Optimization for neural networks Prof. J.C. Kao, UCLA Optimization for neural networks We previously introduced the principle of gradient descent. Now we will discuss specific modifications we make

### Conjugate Directions for Stochastic Gradient Descent

Conjugate Directions for Stochastic Gradient Descent Nicol N Schraudolph Thore Graepel Institute of Computational Science ETH Zürich, Switzerland {schraudo,graepel}@infethzch Abstract The method of conjugate

### Comparison of methods for finding saddle points without knowledge of the final states

JOURNAL OF CHEMICAL PHYSICS VOLUME 121, NUMBER 20 22 NOVEMBER 2004 Comparison of methods for finding saddle points without knowledge of the final states R. A. Olsen and G. J. Kroes Leiden Institute of

### EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Taylor s Theorem Can often approximate a function by a polynomial The error in the approximation

### 11 More Regression; Newton s Method; ROC Curves

More Regression; Newton s Method; ROC Curves 59 11 More Regression; Newton s Method; ROC Curves LEAST-SQUARES POLYNOMIAL REGRESSION Replace each X i with feature vector e.g. (X i ) = [X 2 i1 X i1 X i2

### Mobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti

Mobile Robotics 1 A Compact Course on Linear Algebra Giorgio Grisetti SA-1 Vectors Arrays of numbers They represent a point in a n dimensional space 2 Vectors: Scalar Product Scalar-Vector Product Changes

### Optimization Methods

Optimization Methods Categorization of Optimization Problems Continuous Optimization Discrete Optimization Combinatorial Optimization Variational Optimization Common Optimization Concepts in Computer Vision

### ECE 595, Section 10 Numerical Simulations Lecture 7: Optimization and Eigenvalues. Prof. Peter Bermel January 23, 2013

ECE 595, Section 10 Numerical Simulations Lecture 7: Optimization and Eigenvalues Prof. Peter Bermel January 23, 2013 Outline Recap from Friday Optimization Methods Brent s Method Golden Section Search

### AM 205: lecture 18. Last time: optimization methods Today: conditions for optimality

AM 205: lecture 18 Last time: optimization methods Today: conditions for optimality Existence of Global Minimum For example: f (x, y) = x 2 + y 2 is coercive on R 2 (global min. at (0, 0)) f (x) = x 3

### The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

### Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted

### Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted

### Design and Optimization of Energy Systems Prof. C. Balaji Department of Mechanical Engineering Indian Institute of Technology, Madras

Design and Optimization of Energy Systems Prof. C. Balaji Department of Mechanical Engineering Indian Institute of Technology, Madras Lecture - 09 Newton-Raphson Method Contd We will continue with our

### Lecture 7: Minimization or maximization of functions (Recipes Chapter 10)

Lecture 7: Minimization or maximization of functions (Recipes Chapter 10) Actively studied subject for several reasons: Commonly encountered problem: e.g. Hamilton s and Lagrange s principles, economics

### Geometry Optimisation

Geometry Optimisation Matt Probert Condensed Matter Dynamics Group Department of Physics, University of York, UK http://www.cmt.york.ac.uk/cmd http://www.castep.org Motivation Overview of Talk Background

### Tangent spaces, normals and extrema

Chapter 3 Tangent spaces, normals and extrema If S is a surface in 3-space, with a point a S where S looks smooth, i.e., without any fold or cusp or self-crossing, we can intuitively define the tangent

### Day 3 Lecture 3. Optimizing deep networks

Day 3 Lecture 3 Optimizing deep networks Convex optimization A function is convex if for all α [0,1]: f(x) Tangent line Examples Quadratics 2-norms Properties Local minimum is global minimum x Gradient

### Regression with Numerical Optimization. Logistic

CSG220 Machine Learning Fall 2008 Regression with Numerical Optimization. Logistic regression Regression with Numerical Optimization. Logistic regression based on a document by Andrew Ng October 3, 204

### 1 Numerical optimization

Contents 1 Numerical optimization 5 1.1 Optimization of single-variable functions............ 5 1.1.1 Golden Section Search................... 6 1.1. Fibonacci Search...................... 8 1. Algorithms

### Statistics 580 Optimization Methods

Statistics 580 Optimization Methods Introduction Let fx be a given real-valued function on R p. The general optimization problem is to find an x ɛ R p at which fx attain a maximum or a minimum. It is of

Quadratic Programming Outline Linearly constrained minimization Linear equality constraints Linear inequality constraints Quadratic objective function 2 SideBar: Matrix Spaces Four fundamental subspaces

### Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 2018 3 Lecture 3 3.1 General remarks March 4, 2018 This

### Scientific Computing: Optimization

Scientific Computing: Optimization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 March 8th, 2011 A. Donev (Courant Institute) Lecture

### Lecture 8 Optimization

4/9/015 Lecture 8 Optimization EE 4386/5301 Computational Methods in EE Spring 015 Optimization 1 Outline Introduction 1D Optimization Parabolic interpolation Golden section search Newton s method Multidimensional

### Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

### Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method

Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 3 Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 2 3.1. Gradient method Classical gradient method: to minimize a differentiable convex

### LECTURE 22: SWARM INTELLIGENCE 3 / CLASSICAL OPTIMIZATION

15-382 COLLECTIVE INTELLIGENCE - S19 LECTURE 22: SWARM INTELLIGENCE 3 / CLASSICAL OPTIMIZATION TEACHER: GIANNI A. DI CARO WHAT IF WE HAVE ONE SINGLE AGENT PSO leverages the presence of a swarm: the outcome

### A Study of the Thermal Properties of a One. Dimensional Lennard-Jones System

A Study of the Thermal Properties of a One Dimensional Lennard-Jones System Abstract In this study, the behavior of a one dimensional (1D) Lennard-Jones (LJ) system is simulated. As part of this research,

### Geometry optimization

Geometry optimization Trygve Helgaker Centre for Theoretical and Computational Chemistry Department of Chemistry, University of Oslo, Norway European Summer School in Quantum Chemistry (ESQC) 211 Torre

### Lecture 5: Logistic Regression. Neural Networks

Lecture 5: Logistic Regression. Neural Networks Logistic regression Comparison with generative models Feed-forward neural networks Backpropagation Tricks for training neural networks COMP-652, Lecture

### Analysis of Crossover Operators for Cluster Geometry Optimization

Analysis of Crossover Operators for Cluster Geometry Optimization Francisco B. Pereira Instituto Superior de Engenharia de Coimbra Portugal Abstract We study the effectiveness of different crossover operators

### Introduction to Optimization

Introduction to Optimization Konstantin Tretyakov (kt@ut.ee) MTAT.03.227 Machine Learning So far Machine learning is important and interesting The general concept: Fitting models to data So far Machine

### Landscape Approach to Glass Transition and Relaxation. Lecture # 4, April 1 (last of four lectures)

Landscape Approach to Glass Transition and Relaxation Lecture # 4, April 1 (last of four lectures) Relaxation in the glassy state Instructor: Prabhat Gupta The Ohio State University (gupta.3@osu.edu) Review

### Topic 8c Multi Variable Optimization

Course Instructor Dr. Raymond C. Rumpf Office: A 337 Phone: (915) 747 6958 E Mail: rcrumpf@utep.edu Topic 8c Multi Variable Optimization EE 4386/5301 Computational Methods in EE Outline Mathematical Preliminaries

### Lattice dynamics. Javier Junquera. Philippe Ghosez. Andrei Postnikov

Lattice dynamics Javier Junquera Philippe Ghosez Andrei Postnikov Warm up: a little bit of notation Greek characters ( ) refer to atoms within the unit cell Latin characters ( ) refer to the different

### Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Spring Semester 2006 Christopher J. Cramer. Lecture 9, February 8, 2006

Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Spring Semester 2006 Christopher J. Cramer Lecture 9, February 8, 2006 The Harmonic Oscillator Consider a diatomic molecule. Such a molecule

### Introduction to Optimization

Introduction to Optimization Konstantin Tretyakov (kt@ut.ee) MTAT.03.227 Machine Learning So far Machine learning is important and interesting The general concept: Fitting models to data So far Machine

### AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods

AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality

Chapter 8 Gradient Methods An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Introduction Recall that a level set of a function is the set of points satisfying for some constant. Thus, a point

### x k+1 = x k + α k p k (13.1)

13 Gradient Descent Methods Lab Objective: Iterative optimization methods choose a search direction and a step size at each iteration One simple choice for the search direction is the negative gradient,

### Convex Optimization CMU-10725

Convex Optimization CMU-10725 Quasi Newton Methods Barnabás Póczos & Ryan Tibshirani Quasi Newton Methods 2 Outline Modified Newton Method Rank one correction of the inverse Rank two correction of the

### Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23

Optimization: Nonlinear Optimization without Constraints Nonlinear Optimization without Constraints 1 / 23 Nonlinear optimization without constraints Unconstrained minimization min x f(x) where f(x) is

### Multi-Ensemble Markov Models and TRAM. Fabian Paul 21-Feb-2018

Multi-Ensemble Markov Models and TRAM Fabian Paul 21-Feb-2018 Outline Free energies Simulation types Boltzmann reweighting Umbrella sampling multi-temperature simulation accelerated MD Analysis methods

### 1. Introductory Examples

1. Introductory Examples We introduce the concept of the deterministic and stochastic simulation methods. Two problems are provided to explain the methods: the percolation problem, providing an example

### On fast trust region methods for quadratic models with linear constraints. M.J.D. Powell

DAMTP 2014/NA02 On fast trust region methods for quadratic models with linear constraints M.J.D. Powell Abstract: Quadratic models Q k (x), x R n, of the objective function F (x), x R n, are used by many

### Analysis of the Lennard-Jones-38 stochastic network

Analysis of the Lennard-Jones-38 stochastic network Maria Cameron Joint work with E. Vanden-Eijnden Lennard-Jones clusters Pair potential: V(r) = 4e(r -12 - r -6 ) 1. Wales, D. J., Energy landscapes: calculating

### Simulations with MM Force Fields. Monte Carlo (MC) and Molecular Dynamics (MD) Video II.vi

Simulations with MM Force Fields Monte Carlo (MC) and Molecular Dynamics (MD) Video II.vi Some slides taken with permission from Howard R. Mayne Department of Chemistry University of New Hampshire Walking

### A simple technique to estimate partition functions and equilibrium constants from Monte Carlo simulations

A simple technique to estimate partition functions and equilibrium constants from Monte Carlo simulations Michal Vieth Department of Chemistry, The Scripps Research Institute, 10666 N. Torrey Pines Road,

### Numerical Methods I Solving Nonlinear Equations

Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)

### Static unconstrained optimization

Static unconstrained optimization 2 In unconstrained optimization an objective function is minimized without any additional restriction on the decision variables, i.e. min f(x) x X ad (2.) with X ad R

### Chapter 4. Unconstrained optimization

Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file

### 8 Numerical methods for unconstrained problems

8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields

### Department of Chemical Engineering University of California, Santa Barbara Spring Exercise 2. Due: Thursday, 4/19/09

Department of Chemical Engineering ChE 210D University of California, Santa Barbara Spring 2012 Exercise 2 Due: Thursday, 4/19/09 Objective: To learn how to compile Fortran libraries for Python, and to

### QUASI-EQUILIBRIUM MONTE-CARLO: OFF-LATTICE KINETIC MONTE CARLO SIMULATION OF HETEROEPITAXY WITHOUT SADDLE POINTS

QUASI-EQUILIBRIUM MONTE-CARLO: OFF-LATTICE KINETIC MONTE CARLO SIMULATION OF HETEROEPITAXY WITHOUT SADDLE POINTS Henry A. Boateng University of Michigan, Ann Arbor Joint work with Tim Schulze and Peter

### Line Search Algorithms

Lab 1 Line Search Algorithms Investigate various Line-Search algorithms for numerical opti- Lab Objective: mization. Overview of Line Search Algorithms Imagine you are out hiking on a mountain, and you