Problem Set 5: Iterative methods with generalized least squares

Size: px
Start display at page:

Download "Problem Set 5: Iterative methods with generalized least squares"

Transcription

1 Problem Set 5: Iterative methods with generalized least squares GEOS 627: Inverse Problems and Parameter Estimation, Carl Tape Assigned: February 20, 2017 Due: March 6, 2017 Last compiled: July 7, 2017 Overview and instructions 1. Reading: Section 6.22 of Tarantola (2005) 2. Problem 1. In HW4 you explored Gaussian random fields with prescribed covariance (exponential or Gaussian) for a 1D spatial field. This problem deals with 2D Gaussian random fields. Your goal is to understand how these fields are generated, and to get some intuitive sense for how certain parameters change the appearance of the 2D fields. The Matlab scripts for computing Gaussian random fields in the frequency domain were provided by Miranda Holmes, who wrote them as part of the study Bühler and Holmes- Cerfon (2009). She also wrote a useful set of notes, Generating stationary Gaussian random fields, which is available as ft_summary.pdf in the same directory containing the Matlab scripts (~/inv2017/util_grf/). Why do we care about the statistical characteristics of spatial random fields? Because real systems have complex variations that we would like to quantify in simpler terms. Gneiting et al. (2012) examined several different covariance functions, including the Matern family, which includes the Gaussian and exponential functions, and they illustrate their technique with line transects of arctic sea ice. 3. Problem 2. Iterative methods are needed for nonlinear problems, which are characterized by a nonlinear forward model g(m). In many problems it is not computationally feasible to evaluate the misfit function dozens of times, let alone millions (or billions) of times needed to cover the M-dimensional model space. The basic strategy is to evaluate the misfit function at one point, then use the gradient (and Hessian) at that point to guide the choice of the next point. This problem will build upon the lab exercie lab_iter.pdf. 4. Problem 3. Here we revisit Tarantola (2005), Problem 7-1. See HW3 for background. 1

2 Problem 1 (3.0). 2D Gaussian random fields In this problem, a sample is a 2D spatial field that is n x n y. 1. (0.2) Make a copy of the template code: cp covrand2d_template.m covrand2d.m Run covrand2d.m, which will generate 1000 samples of a 2D Gaussian random field with prescribed covariance; eight samples are plotted. Compute and plot the sample mean µ 1000 with a colorbar. Throughout this problem, when plotting samples, use the same plotting commands 1 : axis equal, axis(ax1aex), caxis(3*sigma*[-1 1]) 2. (0.3) Open xy2distance.m 2, set idisplay=1, then run the example at the bottom. (a) (0.0) Explain how the points are ordered within the distance index matrix id. (b) (0.1) What kind of matrix structure does id have? (Be as specific as possible.) (c) (0.1) How can you compute the actual distances between points, D (or D)? (Recognize the difference between the distance matrix D and the index distance matrix.) (d) (0.1) What is the maximum distance between two points in the example grid? Note that you need to consider x, which is not represented in the plotted example grid. After you are done, be sure to reset idisplay=0. 3. (0.0) What is the maximum distance between two points in the default grid in covrand.m? 4. (0.5) Compute the sample covariance matrix C 1000, which we label as Csamp. (a) (0.2) Include a plot of C 1000 figure; imagesc(csamp); axis equal, axis tight (b) (0.1) Plot the points of C 1000 vs D. We denote this as C 1000 (d). (c) (0.1) Superimpose the covariance function, C(d). Hint: plot(id*dx,c, r. ) is one way to plot C(d). (d) (0.1) Show that the length scale is consistent with the input value L (Matlab variable L) (hint: see HW4 solutions). 5. (0.0) Now set ifourier=1 and idouble=1 and convince yourself that the FFT method gives the same result. Check that your scatterplot of estimated C 1000 (d) is about the same. 6. (0.3) The Cholesky decomposition is extremely unstable. In order to consider denser grids, or covariance functions with large L length scales, we need to use Fourier methods (ifourier=1). (a) Explain what happens in the code if idouble=0 instead of idouble=1. 1 If you are not using Matlab, then they key points when plotting are to not distort the shape of the samples and to use a color scale that ranges between 3σ and +3σ. 2 This is in the subfolder util grf, which will be in your Matlab path (type open xy2distance) if you ran covrand2d.m in the previous problem. 2

3 (b) Include a plot of C 1000 for idouble=0. (c) Explain the impact of changing idouble on the samples. (After you are done, be sure to reset idouble=1.) From here on out, we will use idouble=1. 7. (0.2) Compute the mean and standard deviation of each sample from the set of 1000 samples. Plot the means and standard deviations as two histograms. 8. (1.0) Generate a Gaussian random field with (a) Gaussian covariance, (b) exponential covariance, and (c) circular covariance. In each case, use the same Gaussian random vector (note: this requires coding). Keep all other parameters fixed (as defaults), and use nx = 2^7, ichol=0, ifourier=1, idouble=1. (a) Show the GRFs in a 3 1 subplot figure. In all plots use the same the color range caxis(3*sigma*[-1 1]). (b) Show your modified lines of code. Hint: Use the template code provided at the end of covrand2d.m and see grf2.m. 9. (0.5) Choose a covariance function (pick any icov), then plot six Gaussian random fields, each with a different length scale L. Use the same Gaussian random vector to generate each GRF. Include a 3 2 subplot figure showing your GRFs. Problem 2 (4.0). Implementation of iterative methods See the lab exercise on the quasi-newton method (lab_iter.pdf). There you were asked to adapt genlsq_template.m to write a functioning version of the quasi-newton algorithm for the 4-parameter epicenter problem. Moving forward, it is important that your code is correct. Start with genlsq_sol_template.m cp genlsq_sol_template.m genlsq_iter.m Alternatively you can use your genlsq.m script, but, whichever you use, check your results for the quasi-newton method with those listed in Table 1. In this problem you will replace the quasi-newton method with three other methods. You will not need to touch the code associated with the forward model (forward_epicenter.m). 1. (1.5) Implement the steepest descent method (Eq ). Use Eq for µ n. Use 8 iterations (niter=8). Include the following: (a) your code (b) a plot of the misfit reduction with iteration (note: this plot is produced by default) (c) a plot showing epicenter samples of the prior and posterior models (with m target and m initial ) (note: this plot is produced by default) (d) the posterior model (in Table 1; list numbers to precision) 2. (1.5) Repeat Problem 2-1 for the conjugate gradient method (Eq ). Use Eq for µ n. 3

4 Use Eq for α n. Use F 0 = I such that λ n = F 0 γ n = γ n. Note that the search direction is initialized as φ 0 = λ 0 (= γ 0 ). 3. (0.5) Repeat Problem 2-1 for the variable metric method (Section ). Use Eq for µ n. Use F 0 = I. Use Eq for F n+1, but note that there is a typo: there should be no transpose on the last δγ term in the denominator. Hint: Write the equations in non-hat notation, such as in Eq For example, note that ˆFˆγ = Fγ. 4. (0.5) (a) (0.4) Compare and contrast these three methods (see Tarantola, 2005). NOTE: This problem can be answered even if your implementations in 2-1, 2-2, 2-3 were unsuccessful. (b) (0.1) Compare the performance of each method for our problem. Problem 3 (3.0). Revisiting Tarantola (2005), Problem 7-1 Your goal is to implement Tarantola (2005), Problem 7-1. Prepare for this example by copying two files: cp genlsq_sol_template.m genlsq_crescent.m cp forward_epicenter.m forward_epicenter_crescent.m Run genlsq_crescent.m and check your quasi-newton results with those listed in Table 1. The file forward_epicenter_crescent.m will serve as a template for this problem. genlsq_crescent.m contains the iterative inverse problem, and it also calls the forward model forward_epicenter.m. In genlsq_crescent.m, replace the call to forward_epicenter.m with forward_epicenter_crescent.m. 1. (2.5) Adapt forward_epicenter_crescent.m for Problem 7-1 of Tarantola (2005) (HW3). In addition to using the values used in Problem 7-1, make the following additional choices: Define m prior to be the center of the plotting grid used for HW3. The code you want to use is this: % range of model space (Tarantola Figure 7.1) xmin = 0; xmax = 22; ymin = -2; ymax = 30; xcen = (xmax+xmin)/2; ycen = (ymin+ymax)/2; % prior model mprior = [ xcen ycen ] ; 4

5 Use the same σ values for the prior epicenter as inforward_epicenter.m (σ xs = σ ys = 10 km). This prior model is chosen as an analog for the perspective that the epicenter could be anywhere within a large region (such as the plotting grid). m initial = (15,20). m target = (15,5). (This is probably what Tarantola used, though we can t be certain.) For the case of fixed data errors, set tobs = [ ] ; eobs = tobs - dtarget; This allows us to remove the errors added to the arrival time data listed in Tarantola. (It gets added back in as dobs = dtarget + eobs, which is tobs.) axepi = [xmin xmax ymin ymax]; Solve the problem using the quasi-newton method with eight iterations. (a) Include figures showing (a) the misfit reduction; (b) the prior and posterior samples, along with the initial model. (Note that these figures are automatically generated.) (b) List the solution after eight iterations: m post, C post, and the correlation matrix ρ post. Complete Table 2. (c) How many iterations are needed for convergence? 2. (0.5) See Aster et al. (2013, p ) for how to compute confidence regions. The key concept is that the inequality (m m post ) T C 1 post (m m post) 2 describes the interior region of an M-dimensional ellipsoid (in our case, M = 2). For example, 2 can be chosen to represent the boundary of the 95% confidence region. (a) (0.1) Use eig to compute the eigen-decomposition of C post. Compute the quantity λ max /λ min. (b) (0.1) Use delta2 = chi2inv(0.95,2) to compute 2. Hint: See ex_2_1_1_uaf.m for an example of using chi2inv. (c) (0.1) Compute the lengths of the semi-major axis and semi-minor axis of the ellipse, where the length of the kth axis is λ k. (d) (0.2) Plot the confidence region using plot_ellipse. (You should see agreement between the locations of your samples of C post and the ellipse.) Also include the ellipse axes in your plot (plus m post, m initial, m target, samples of prior and posterior, etc). NOTE: Even if you did not successfully implement forward_epicenter_crescent.m, you can still do the confidence region for the epicenter associated with the forward problem of forward_epicenter.m Problem Approximately how much time outside of class and lab time did you spend on this problem set? Feel free to suggest improvements here. 5

6 References Aster, R. C., B. Borchers, and C. H. Thurber (2013), Parameter Estimation and Inverse Problems, 2 ed., Elsevier, Waltham, Mass., USA. Bühler, O., and M. Holmes-Cerfon (2009), Particle dispersion by random waves in rotating shallow water, J. Fluid Mech., 638, Gneiting, T., H. Sevcikova, and D. P. Percival (2012), Estimators of fractal dimension: Assessing the roughness of time series and spatial data, Statistical Science, 27(2), Tarantola, A. (2005), Inverse Problem Theory and Methods for Model Parameter Estimation, SIAM, Philadelphia, Penn., USA. Table 1: Summary of results for the four iterative methods. Posterior models are listed for the eighth iteration: m post = m 8. QN = quasi-newton; SD = steepest descent; CG = conjugate gradient; VM = variable metric. List numbers to precision. prior initial target QN SD CG VM x s, km y s, km t s, s v Table 2: Epicenter problem after eight iterations with the quasi-newton method. List numbers to precision. x s, km y s, km prior initial target QN 6

FALL 2018 MATH 4211/6211 Optimization Homework 4

FALL 2018 MATH 4211/6211 Optimization Homework 4 FALL 2018 MATH 4211/6211 Optimization Homework 4 This homework assignment is open to textbook, reference books, slides, and online resources, excluding any direct solution to the problem (such as solution

More information

Optimization method examples in Matlab Carl Tape, Qinya Liu, Albert Tarantola July 2, The forward problem 3

Optimization method examples in Matlab Carl Tape, Qinya Liu, Albert Tarantola July 2, The forward problem 3 Optimization method examples in Matlab Carl Tape, Qinya Liu, Albert Tarantola July 2, 27 This file : /home/carltape/latex/notes/tromp/optimization_guide.tex Matlab code : /opt/seismo-util/source/optim_matlab/optimization.m

More information

Steepest descent minimization applied to a simple seismic inversion

Steepest descent minimization applied to a simple seismic inversion Steepest descent minimization applied to a simple seismic inversion Sylvain Barbot Friday, December 10, 2004 1 Problem Definition The problem one tries to address is the estimation of the location of a

More information

2D Plotting with Matlab

2D Plotting with Matlab GEEN 1300 Introduction to Engineering Computing Class Meeting #22 Monday, Nov. 9 th Engineering Computing and Problem Solving with Matlab 2-D plotting with Matlab Script files User-defined functions Matlab

More information

SOLUTIONS to Exercises from Optimization

SOLUTIONS to Exercises from Optimization SOLUTIONS to Exercises from Optimization. Use the bisection method to find the root correct to 6 decimal places: 3x 3 + x 2 = x + 5 SOLUTION: For the root finding algorithm, we need to rewrite the equation

More information

Uncertainty quantification for Wavefield Reconstruction Inversion

Uncertainty quantification for Wavefield Reconstruction Inversion Uncertainty quantification for Wavefield Reconstruction Inversion Zhilong Fang *, Chia Ying Lee, Curt Da Silva *, Felix J. Herrmann *, and Rachel Kuske * Seismic Laboratory for Imaging and Modeling (SLIM),

More information

HOMEWORK 3: Phase portraits for Mathmatical Models, Analysis and Simulation, Fall 2010 Report due Mon Oct 4, Maximum score 7.0 pts.

HOMEWORK 3: Phase portraits for Mathmatical Models, Analysis and Simulation, Fall 2010 Report due Mon Oct 4, Maximum score 7.0 pts. HOMEWORK 3: Phase portraits for Mathmatical Models, Analysis and Simulation, Fall 2010 Report due Mon Oct 4, 2010. Maximum score 7.0 pts. Three problems are to be solved in this homework assignment. The

More information

JMBC Computational Fluid Dynamics I Exercises by A.E.P. Veldman

JMBC Computational Fluid Dynamics I Exercises by A.E.P. Veldman JMBC Computational Fluid Dynamics I Exercises by A.E.P. Veldman The exercises will be carried out on PC s in the practicum rooms. Several (Matlab and Fortran) files are required. How these can be obtained

More information

EE3210 Lab 3: Periodic Signal Representation by Fourier Series

EE3210 Lab 3: Periodic Signal Representation by Fourier Series City University of Hong Kong Department of Electronic Engineering EE321 Lab 3: Periodic Signal Representation by Fourier Series Prelab: Read the Background section. Complete Section 2.2(b), which asks

More information

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method. Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization

More information

Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

More information

Numerical Optimization Prof. Shirish K. Shevade Department of Computer Science and Automation Indian Institute of Science, Bangalore

Numerical Optimization Prof. Shirish K. Shevade Department of Computer Science and Automation Indian Institute of Science, Bangalore Numerical Optimization Prof. Shirish K. Shevade Department of Computer Science and Automation Indian Institute of Science, Bangalore Lecture - 13 Steepest Descent Method Hello, welcome back to this series

More information

HOMEWORK #4: LOGISTIC REGRESSION

HOMEWORK #4: LOGISTIC REGRESSION HOMEWORK #4: LOGISTIC REGRESSION Probabilistic Learning: Theory and Algorithms CS 274A, Winter 2019 Due: 11am Monday, February 25th, 2019 Submit scan of plots/written responses to Gradebook; submit your

More information

Nonlinear Optimization for Optimal Control

Nonlinear Optimization for Optimal Control Nonlinear Optimization for Optimal Control Pieter Abbeel UC Berkeley EECS Many slides and figures adapted from Stephen Boyd [optional] Boyd and Vandenberghe, Convex Optimization, Chapters 9 11 [optional]

More information

7.2 Steepest Descent and Preconditioning

7.2 Steepest Descent and Preconditioning 7.2 Steepest Descent and Preconditioning Descent methods are a broad class of iterative methods for finding solutions of the linear system Ax = b for symmetric positive definite matrix A R n n. Consider

More information

Review for Exam 2 Ben Wang and Mark Styczynski

Review for Exam 2 Ben Wang and Mark Styczynski Review for Exam Ben Wang and Mark Styczynski This is a rough approximation of what we went over in the review session. This is actually more detailed in portions than what we went over. Also, please note

More information

The Inversion Problem: solving parameters inversion and assimilation problems

The Inversion Problem: solving parameters inversion and assimilation problems The Inversion Problem: solving parameters inversion and assimilation problems UE Numerical Methods Workshop Romain Brossier romain.brossier@univ-grenoble-alpes.fr ISTerre, Univ. Grenoble Alpes Master 08/09/2016

More information

Numerical computation II. Reprojection error Bundle adjustment Family of Newtonʼs methods Statistical background Maximum likelihood estimation

Numerical computation II. Reprojection error Bundle adjustment Family of Newtonʼs methods Statistical background Maximum likelihood estimation Numerical computation II Reprojection error Bundle adjustment Family of Newtonʼs methods Statistical background Maximum likelihood estimation Reprojection error Reprojection error = Distance between the

More information

SV6: Polynomial Regression and Neural Networks

SV6: Polynomial Regression and Neural Networks Signal and Information Processing Laboratory Institut für Signal- und Informationsverarbeitung Fachpraktikum Signalverarbeitung SV6: Polynomial Regression and Neural Networks 1 Introduction Consider the

More information

Foundations for Functions

Foundations for Functions Activity: TEKS: Overview: Materials: Regression Exploration (A.2) Foundations for functions. The student uses the properties and attributes of functions. The student is expected to: (D) collect and organize

More information

User Guide for Hermir version 0.9: Toolbox for Synthesis of Multivariate Stationary Gaussian and non-gaussian Series

User Guide for Hermir version 0.9: Toolbox for Synthesis of Multivariate Stationary Gaussian and non-gaussian Series User Guide for Hermir version 0.9: Toolbox for Synthesis of Multivariate Stationary Gaussian and non-gaussian Series Hannes Helgason, Vladas Pipiras, and Patrice Abry June 2, 2011 Contents 1 Organization

More information

Some definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts

Some definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts Some definitions Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization M. M. Sussman sussmanm@math.pitt.edu Office Hours: MW 1:45PM-2:45PM, Thack 622 A matrix A is SPD (Symmetric

More information

Math 148. Polynomial Graphs

Math 148. Polynomial Graphs Math 148 Lab 1 Polynomial Graphs Due: Monday Wednesday, April April 10 5 Directions: Work out each problem on a separate sheet of paper, and write your answers on the answer sheet provided. Submit the

More information

Numerical Optimization

Numerical Optimization Numerical Optimization Unit 2: Multivariable optimization problems Che-Rung Lee Scribe: February 28, 2011 (UNIT 2) Numerical Optimization February 28, 2011 1 / 17 Partial derivative of a two variable function

More information

Probabilistic Machine Learning. Industrial AI Lab.

Probabilistic Machine Learning. Industrial AI Lab. Probabilistic Machine Learning Industrial AI Lab. Probabilistic Linear Regression Outline Probabilistic Classification Probabilistic Clustering Probabilistic Dimension Reduction 2 Probabilistic Linear

More information

Lesson 5 Practice Problems

Lesson 5 Practice Problems Name: Date: Lesson 5 Section 5.1: Linear Functions vs. Exponential Functions 1. Complete the table below. Function Linear or Exponential? Linear: Increasing or Decreasing? Exponential: Growth or Decay?

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week

More information

MA/OR/ST 706: Nonlinear Programming Midterm Exam Instructor: Dr. Kartik Sivaramakrishnan INSTRUCTIONS

MA/OR/ST 706: Nonlinear Programming Midterm Exam Instructor: Dr. Kartik Sivaramakrishnan INSTRUCTIONS MA/OR/ST 706: Nonlinear Programming Midterm Exam Instructor: Dr. Kartik Sivaramakrishnan INSTRUCTIONS 1. Please write your name and student number clearly on the front page of the exam. 2. The exam is

More information

= (G T G) 1 G T d. m L2

= (G T G) 1 G T d. m L2 The importance of the Vp/Vs ratio in determining the error propagation and the resolution in linear AVA inversion M. Aleardi, A. Mazzotti Earth Sciences Department, University of Pisa, Italy Introduction.

More information

Approximate- vs. full-hessian in FWI: 1D analytical and numerical experiments

Approximate- vs. full-hessian in FWI: 1D analytical and numerical experiments Approximate- vs. full-hessian in FWI: 1D analytical and numerical experiments Raul Cova and Kris Innanen ABSTRACT Feasibility of using Full Waveform Inversion (FWI) to build velocity models has been increasing

More information

Course on Inverse Problems

Course on Inverse Problems California Institute of Technology Division of Geological and Planetary Sciences March 26 - May 25, 2007 Course on Inverse Problems Albert Tarantola Institut de Physique du Globe de Paris Lesson XVI CONCLUSION

More information

DOING PHYSICS WITH MATLAB ELECTRIC FIELD AND ELECTRIC POTENTIAL: POISSON S EQUATION

DOING PHYSICS WITH MATLAB ELECTRIC FIELD AND ELECTRIC POTENTIAL: POISSON S EQUATION DOING PHYSICS WITH MATLAB ELECTRIC FIELD AND ELECTRIC POTENTIAL: POISSON S EQUATION Ian Cooper School of Physics, University of Sydney ian.cooper@sydney.edu.au DOWNLOAD DIRECTORY FOR MATLAB SCRIPTS For

More information

Lab 2: Photon Counting with a Photomultiplier Tube

Lab 2: Photon Counting with a Photomultiplier Tube Lab 2: Photon Counting with a Photomultiplier Tube 1 Introduction 1.1 Goals In this lab, you will investigate properties of light using a photomultiplier tube (PMT). You will assess the quantitative measurements

More information

Course on Inverse Problems Albert Tarantola

Course on Inverse Problems Albert Tarantola California Institute of Technology Division of Geological and Planetary Sciences March 26 - May 25, 27 Course on Inverse Problems Albert Tarantola Institut de Physique du Globe de Paris CONCLUSION OF THE

More information

Fall 2017 Qualifier Exam: OPTIMIZATION. September 18, 2017

Fall 2017 Qualifier Exam: OPTIMIZATION. September 18, 2017 Fall 2017 Qualifier Exam: OPTIMIZATION September 18, 2017 GENERAL INSTRUCTIONS: 1 Answer each question in a separate book 2 Indicate on the cover of each book the area of the exam, your code number, and

More information

Lecture Notes: Geometric Considerations in Unconstrained Optimization

Lecture Notes: Geometric Considerations in Unconstrained Optimization Lecture Notes: Geometric Considerations in Unconstrained Optimization James T. Allison February 15, 2006 The primary objectives of this lecture on unconstrained optimization are to: Establish connections

More information

1.) Suppose the graph of f(x) looks like this (each tick mark denotes 1 unit). x y

1.) Suppose the graph of f(x) looks like this (each tick mark denotes 1 unit). x y College Algebra Summer 2014 Exam File Exam #1 1.) Suppose the graph of f(x) looks like this (each tick mark denotes 1 unit). Graph g(x) = -0.5 f(x + 1) - 3 2.) Consider the following table of values. x

More information

Motion II. Goals and Introduction

Motion II. Goals and Introduction Motion II Goals and Introduction As you have probably already seen in lecture or homework, and if you ve performed the experiment Motion I, it is important to develop a strong understanding of how to model

More information

DOING PHYSICS WITH MATLAB. ELECTRIC FIELD AND ELECTRIC POTENTIAL: POISSON S and LAPLACES S EQUATIONS

DOING PHYSICS WITH MATLAB. ELECTRIC FIELD AND ELECTRIC POTENTIAL: POISSON S and LAPLACES S EQUATIONS DOING PHYSICS WITH MATLAB ELECTRIC FIELD AND ELECTRIC POTENTIAL: POISSON S and LAPLACES S EQUATIONS Ian Cooper School of Physics, University of Sydney ian.cooper@sydney.edu.au DOWNLOAD DIRECTORY FOR MATLAB

More information

Clustering. Professor Ameet Talwalkar. Professor Ameet Talwalkar CS260 Machine Learning Algorithms March 8, / 26

Clustering. Professor Ameet Talwalkar. Professor Ameet Talwalkar CS260 Machine Learning Algorithms March 8, / 26 Clustering Professor Ameet Talwalkar Professor Ameet Talwalkar CS26 Machine Learning Algorithms March 8, 217 1 / 26 Outline 1 Administration 2 Review of last lecture 3 Clustering Professor Ameet Talwalkar

More information

LINEAR AND NONLINEAR PROGRAMMING

LINEAR AND NONLINEAR PROGRAMMING LINEAR AND NONLINEAR PROGRAMMING Stephen G. Nash and Ariela Sofer George Mason University The McGraw-Hill Companies, Inc. New York St. Louis San Francisco Auckland Bogota Caracas Lisbon London Madrid Mexico

More information

Computational Physics, Spring Homework Assignment # 9

Computational Physics, Spring Homework Assignment # 9 Cornell University Department of Physics Phys 480/680, Astro 690 April 26, 2007 Computational Physics, Spring 2007 Homework Assignment # 9 (Due Wednesday, April 25 at 5:00pm) Agenda and readings: Goal:

More information

Computational Foundations of Cognitive Science

Computational Foundations of Cognitive Science Computational Foundations of Cognitive Science Lecture 14: Inverses and Eigenvectors in Matlab; Plotting and Graphics Frank Keller School of Informatics University of Edinburgh keller@inf.ed.ac.uk February

More information

Lab E3: The Wheatstone Bridge

Lab E3: The Wheatstone Bridge E3.1 Lab E3: The Wheatstone Bridge Introduction The Wheatstone bridge is a circuit used to compare an unknown resistance with a known resistance. The bridge is commonly used in control circuits. For instance,

More information

Towards an Optimal Noise Versus Resolution Trade-off in Wind Scatterometry

Towards an Optimal Noise Versus Resolution Trade-off in Wind Scatterometry Towards an Optimal Noise Versus Resolution Trade-off in Wind Scatterometry Brent Williams Jet Propulsion Lab, California Institute of Technology IOWVST Meeting Utrecht Netherlands June 12, 2012 Copyright

More information

MATH0328: Numerical Linear Algebra Homework 3 SOLUTIONS

MATH0328: Numerical Linear Algebra Homework 3 SOLUTIONS MATH038: Numerical Linear Algebra Homework 3 SOLUTIONS Due Wednesday, March 8 Instructions Complete the following problems. Show all work. Only use Matlab on problems that are marked with (MATLAB). Include

More information

Math 7824 Spring 2010 Numerical solution of partial differential equations Classroom notes and homework

Math 7824 Spring 2010 Numerical solution of partial differential equations Classroom notes and homework Math 7824 Spring 2010 Numerical solution of partial differential equations Classroom notes and homework Jan Mandel University of Colorado Denver May 12, 2010 1/20/09: Sec. 1.1, 1.2. Hw 1 due 1/27: problems

More information

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares Robert Bridson October 29, 2008 1 Hessian Problems in Newton Last time we fixed one of plain Newton s problems by introducing line search

More information

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09 Numerical Optimization 1 Working Horse in Computer Vision Variational Methods Shape Analysis Machine Learning Markov Random Fields Geometry Common denominator: optimization problems 2 Overview of Methods

More information

Physics with Matlab and Mathematica Exercise #1 28 Aug 2012

Physics with Matlab and Mathematica Exercise #1 28 Aug 2012 Physics with Matlab and Mathematica Exercise #1 28 Aug 2012 You can work this exercise in either matlab or mathematica. Your choice. A simple harmonic oscillator is constructed from a mass m and a spring

More information

MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year

MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2. Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year MATHEMATICS FOR COMPUTER VISION WEEK 8 OPTIMISATION PART 2 1 Dr Fabio Cuzzolin MSc in Computer Vision Oxford Brookes University Year 2013-14 OUTLINE OF WEEK 8 topics: quadratic optimisation, least squares,

More information

1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?)

1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?) Math 35 Exam Review SOLUTIONS Overview In this third of the course we focused on linear learning algorithms to model data. summarize: To. Background: The SVD and the best basis (questions selected from

More information

Chapter 1 Statistical Inference

Chapter 1 Statistical Inference Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations

More information

Scientific Computing: Optimization

Scientific Computing: Optimization Scientific Computing: Optimization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 March 8th, 2011 A. Donev (Courant Institute) Lecture

More information

Unconstrained Multivariate Optimization

Unconstrained Multivariate Optimization Unconstrained Multivariate Optimization Multivariate optimization means optimization of a scalar function of a several variables: and has the general form: y = () min ( ) where () is a nonlinear scalar-valued

More information

x k+1 = x k + α k p k (13.1)

x k+1 = x k + α k p k (13.1) 13 Gradient Descent Methods Lab Objective: Iterative optimization methods choose a search direction and a step size at each iteration One simple choice for the search direction is the negative gradient,

More information

Direct Current Resistivity Inversion using Various Objective Functions

Direct Current Resistivity Inversion using Various Objective Functions Direct Current Resistivity Inversion using Various Objective Functions Rowan Cockett Department of Earth and Ocean Science University of British Columbia rcockett@eos.ubc.ca Abstract In geophysical applications

More information

Centre for Mathematical Sciences HT 2017 Mathematical Statistics

Centre for Mathematical Sciences HT 2017 Mathematical Statistics Lund University Stationary stochastic processes Centre for Mathematical Sciences HT 2017 Mathematical Statistics Computer exercise 3 in Stationary stochastic processes, HT 17. The purpose of this exercise

More information

Model-building and parameter estimation

Model-building and parameter estimation Luleå University of Technology Johan Carlson Last revision: July 27, 2009 Measurement Technology and Uncertainty Analysis - E7021E MATLAB homework assignment Model-building and parameter estimation Introduction

More information

Vasil Khalidov & Miles Hansard. C.M. Bishop s PRML: Chapter 5; Neural Networks

Vasil Khalidov & Miles Hansard. C.M. Bishop s PRML: Chapter 5; Neural Networks C.M. Bishop s PRML: Chapter 5; Neural Networks Introduction The aim is, as before, to find useful decompositions of the target variable; t(x) = y(x, w) + ɛ(x) (3.7) t(x n ) and x n are the observations,

More information

Adaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL.

Adaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL. Adaptive Filtering Fundamentals of Least Mean Squares with MATLABR Alexander D. Poularikas University of Alabama, Huntsville, AL CRC Press Taylor & Francis Croup Boca Raton London New York CRC Press is

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning First-Order Methods, L1-Regularization, Coordinate Descent Winter 2016 Some images from this lecture are taken from Google Image Search. Admin Room: We ll count final numbers

More information

Neural Network Training

Neural Network Training Neural Network Training Sargur Srihari Topics in Network Training 0. Neural network parameters Probabilistic problem formulation Specifying the activation and error functions for Regression Binary classification

More information

Matrix-Vector Operations

Matrix-Vector Operations Week3 Matrix-Vector Operations 31 Opening Remarks 311 Timmy Two Space View at edx Homework 3111 Click on the below link to open a browser window with the Timmy Two Space exercise This exercise was suggested

More information

Total=75 min. Materials BLM cut into cards BLM

Total=75 min. Materials BLM cut into cards BLM Unit 2: Day 4: All together now! Math Learning Goals: Minds On: 15 Identify functions as polynomial functions. Consolidate understanding of properties of functions that include: linear, Action: 50 quadratic,

More information

Introduction to Scientific Computing

Introduction to Scientific Computing Introduction to Scientific Computing Benson Muite benson.muite@ut.ee http://kodu.ut.ee/ benson https://courses.cs.ut.ee/2018/isc/spring 26 March 2018 [Public Domain,https://commons.wikimedia.org/wiki/File1

More information

Optimization Tutorial 1. Basic Gradient Descent

Optimization Tutorial 1. Basic Gradient Descent E0 270 Machine Learning Jan 16, 2015 Optimization Tutorial 1 Basic Gradient Descent Lecture by Harikrishna Narasimhan Note: This tutorial shall assume background in elementary calculus and linear algebra.

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428

More information

Convex Optimization. Problem set 2. Due Monday April 26th

Convex Optimization. Problem set 2. Due Monday April 26th Convex Optimization Problem set 2 Due Monday April 26th 1 Gradient Decent without Line-search In this problem we will consider gradient descent with predetermined step sizes. That is, instead of determining

More information

Math 409/509 (Spring 2011)

Math 409/509 (Spring 2011) Math 409/509 (Spring 2011) Instructor: Emre Mengi Study Guide for Homework 2 This homework concerns the root-finding problem and line-search algorithms for unconstrained optimization. Please don t hesitate

More information

OPER 627: Nonlinear Optimization Lecture 9: Trust-region methods

OPER 627: Nonlinear Optimization Lecture 9: Trust-region methods OPER 627: Nonlinear Optimization Lecture 9: Trust-region methods Department of Statistical Sciences and Operations Research Virginia Commonwealth University Sept 25, 2013 (Lecture 9) Nonlinear Optimization

More information

1 Inference for binomial proportion (Matlab/Python)

1 Inference for binomial proportion (Matlab/Python) Bayesian data analysis exercises from 2015 1 Inference for binomial proportion (Matlab/Python) Algae status is monitored in 274 sites at Finnish lakes and rivers. The observations for the 2008 algae status

More information

Lec3p1, ORF363/COS323

Lec3p1, ORF363/COS323 Lec3 Page 1 Lec3p1, ORF363/COS323 This lecture: Optimization problems - basic notation and terminology Unconstrained optimization The Fermat-Weber problem Least squares First and second order necessary

More information

Math 307 Learning Goals. March 23, 2010

Math 307 Learning Goals. March 23, 2010 Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent

More information

AMSC/CMSC 466 Problem set 3

AMSC/CMSC 466 Problem set 3 AMSC/CMSC 466 Problem set 3 1. Problem 1 of KC, p180, parts (a), (b) and (c). Do part (a) by hand, with and without pivoting. Use MATLAB to check your answer. Use the command A\b to get the solution, and

More information

Assessing the solution quality of the earthquake location problem

Assessing the solution quality of the earthquake location problem Semesterarbeit HS 2008 Department of Earth Sciences ETH Zurich Assessing the solution quality of the earthquake location problem Sabrina Schönholzer Supervisor: Technical Supervisor: Prof. Dr. E. Kissling

More information

= L 1 # L o = " L o. = L 1! L o

= L 1 # L o =  L o. = L 1! L o GG303 Lab 9 10/20/09 1 STRAIN I Main Topics A General deformation B Homogeneous 2-D strain and the strain ellipse C Homogeneous 3-D strain and the strain ellipsoid (Flinn Diagrams) D Comments on measuring

More information

Experiment A12 Monte Carlo Night! Procedure

Experiment A12 Monte Carlo Night! Procedure Experiment A12 Monte Carlo Night! Procedure Deliverables: checked lab notebook, printed plots with captions Overview In the real world, you will never measure the exact same number every single time. Rather,

More information

Unconstrained Optimization

Unconstrained Optimization 1 / 36 Unconstrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University February 2, 2015 2 / 36 3 / 36 4 / 36 5 / 36 1. preliminaries 1.1 local approximation

More information

Chapter 3 Numerical Methods

Chapter 3 Numerical Methods Chapter 3 Numerical Methods Part 2 3.2 Systems of Equations 3.3 Nonlinear and Constrained Optimization 1 Outline 3.2 Systems of Equations 3.3 Nonlinear and Constrained Optimization Summary 2 Outline 3.2

More information

Computational Finance

Computational Finance Department of Mathematics at University of California, San Diego Computational Finance Optimization Techniques [Lecture 2] Michael Holst January 9, 2017 Contents 1 Optimization Techniques 3 1.1 Examples

More information

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are Quadratic forms We consider the quadratic function f : R 2 R defined by f(x) = 2 xt Ax b T x with x = (x, x 2 ) T, () where A R 2 2 is symmetric and b R 2. We will see that, depending on the eigenvalues

More information

Optimisation and Operations Research

Optimisation and Operations Research Optimisation and Operations Research Lecture 22: Linear Programming Revisited Matthew Roughan http://www.maths.adelaide.edu.au/matthew.roughan/ Lecture_notes/OORII/ School

More information

Instructor (Brad Osgood)

Instructor (Brad Osgood) TheFourierTransformAndItsApplications-Lecture26 Instructor (Brad Osgood): Relax, but no, no, no, the TV is on. It's time to hit the road. Time to rock and roll. We're going to now turn to our last topic

More information

L3: Review of linear algebra and MATLAB

L3: Review of linear algebra and MATLAB L3: Review of linear algebra and MATLAB Vector and matrix notation Vectors Matrices Vector spaces Linear transformations Eigenvalues and eigenvectors MATLAB primer CSCE 666 Pattern Analysis Ricardo Gutierrez-Osuna

More information

Mathematical optimization

Mathematical optimization Optimization Mathematical optimization Determine the best solutions to certain mathematically defined problems that are under constrained determine optimality criteria determine the convergence of the

More information

Intelligent Embedded Systems Uncertainty, Information and Learning Mechanisms (Part 1)

Intelligent Embedded Systems Uncertainty, Information and Learning Mechanisms (Part 1) Advanced Research Intelligent Embedded Systems Uncertainty, Information and Learning Mechanisms (Part 1) Intelligence for Embedded Systems Ph. D. and Master Course Manuel Roveri Politecnico di Milano,

More information

MS&E 318 (CME 338) Large-Scale Numerical Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization 1 Origins Instructor: Michael Saunders Spring 2015 Notes 9: Augmented Lagrangian Methods

More information

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc. Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix

More information

LHS Algebra Pre-Test

LHS Algebra Pre-Test Your Name Teacher Block Grade (please circle): 9 10 11 12 Course level (please circle): Honors Level 1 Instructions LHS Algebra Pre-Test The purpose of this test is to see whether you know Algebra 1 well

More information

Line Search Algorithms

Line Search Algorithms Lab 1 Line Search Algorithms Investigate various Line-Search algorithms for numerical opti- Lab Objective: mization. Overview of Line Search Algorithms Imagine you are out hiking on a mountain, and you

More information

Lab 2 Worksheet. Problems. Problem 1: Geometry and Linear Equations

Lab 2 Worksheet. Problems. Problem 1: Geometry and Linear Equations Lab 2 Worksheet Problems Problem : Geometry and Linear Equations Linear algebra is, first and foremost, the study of systems of linear equations. You are going to encounter linear systems frequently in

More information

2.1 Identifying Patterns

2.1 Identifying Patterns I. Foundations for Functions 2.1 Identifying Patterns: Leaders' Notes 2.1 Identifying Patterns Overview: Objective: s: Materials: Participants represent linear relationships among quantities using concrete

More information

Numerical Optimization Techniques

Numerical Optimization Techniques Numerical Optimization Techniques Léon Bottou NEC Labs America COS 424 3/2/2010 Today s Agenda Goals Representation Capacity Control Operational Considerations Computational Considerations Classification,

More information

IE 5531: Engineering Optimization I

IE 5531: Engineering Optimization I IE 5531: Engineering Optimization I Lecture 15: Nonlinear optimization Prof. John Gunnar Carlsson November 1, 2010 Prof. John Gunnar Carlsson IE 5531: Engineering Optimization I November 1, 2010 1 / 24

More information

Mathematical Tools for Neuroscience (NEU 314) Princeton University, Spring 2016 Jonathan Pillow. Homework 8: Logistic Regression & Information Theory

Mathematical Tools for Neuroscience (NEU 314) Princeton University, Spring 2016 Jonathan Pillow. Homework 8: Logistic Regression & Information Theory Mathematical Tools for Neuroscience (NEU 34) Princeton University, Spring 206 Jonathan Pillow Homework 8: Logistic Regression & Information Theory Due: Tuesday, April 26, 9:59am Optimization Toolbox One

More information

MATH10212 Linear Algebra B Homework Week 5

MATH10212 Linear Algebra B Homework Week 5 MATH Linear Algebra B Homework Week 5 Students are strongly advised to acquire a copy of the Textbook: D C Lay Linear Algebra its Applications Pearson 6 (or other editions) Normally homework assignments

More information

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both

More information

HST-582J/6.555J/16.456J Biomedical Signal and Image Processing Spring Laboratory Project 3 Blind Source Separation: Fetal & Maternal ECG

HST-582J/6.555J/16.456J Biomedical Signal and Image Processing Spring Laboratory Project 3 Blind Source Separation: Fetal & Maternal ECG HST-582J/6.555J/16.456J Biomedical Signal and Image Processing Spring 2009 DUE: 4/16/09 Laboratory Project 3 Blind Source Separation: Fetal & Maternal ECG 1 Introduction Many health-related problems that

More information

GEPL 4500/5500 Lab 1 Geo-rectifying Imagery Due Wednesday March 2, Geo-rectifying an image:

GEPL 4500/5500 Lab 1 Geo-rectifying Imagery Due Wednesday March 2, Geo-rectifying an image: GEPL 4500/5500 Lab 1 Geo-rectifying Imagery Due Wednesday March 2, 2005 Geo-rectifying an image: You will be creating a geo-referenced images in this lab. The goal of this lab is to introduce you to the

More information