A projected Hessian for full waveform inversion

Size: px
Start display at page:

Download "A projected Hessian for full waveform inversion"

Transcription

1 CWP-679 A projected Hessian for full waveform inversion Yong Ma & Dave Hale Center for Wave Phenomena, Colorado School of Mines, Golden, CO 80401, USA (c) Figure 1. Update directions for one iteration of the conjugate gradient method, the image-guided conjugate gradient method, and a quasi-newton method with application of the inverse projected Hessian (c). ABSTRACT A Hessian matrix in full waveform inversion (FWI) is difficult to compute directly because of high computational cost and an especially large memory requirement. Therefore, Newton-like methods are rarely feasible in realistic largesize FWI problems. We modify the BFGS method to use a projected Hessian matrix that reduces both the computational cost and memory required, thereby making a quasinewton solution to FWI feasible. Key words: projected Hessian matrix, BFGS method 1 INTRODUCTION Full waveform inversion (FWI) (Tarantola, 1984; Pratt, 1999) is usually formulated as an optimization problem (Symes, 2008), in which we consider the minimization of a nonlinear objective function E: RN R, min E (m), m RN (1) where the function variable m denotes N parameters, such as seismic velocities, for an earth model. In FWI, the objective function often takes a least-squares form: E (m) = 21 kd F (m) k2, where k.k denotes an L2 norm, d denotes the recorded data, and F is a forward datasynthesizing operator, a nonlinear function of the model m. Various iterative methods (Nocedal and Wright, 2000; Tarantola, 2005) for solving this least-squares optimization problem include the conjugate gradient method, Newton s method, the Gauss-Newton method, and various quasi-newton methods. Compared to the conjugate gradient method, Newton s method and other Newton-like methods generally converge faster in fewer iterations. However, Newton-like methods are rarely used to solve realistic seismic full waveform problems, because for such large problems they are costly. Newtonlike methods, especially Newton s method, are suitable only for solving small- or medium-size optimization problems (Pratt et al., 1998; Sheen et al., 2006; Virieux and Operto, 2009), in which the number N of model parameters ranges from hundreds to thousands. Newton s method must compute the 1st and 2nd order derivatives of the objective function E (m) with respect to model parameters in m. We refer to these two E and the Hesderivatives as the gradient g (m) m 2 E sian H (m) m2, respectively. The Hessian matrix, in particular, consists of large numbers O(N 2 ) of 2nd

2 14 Y. Ma & D. Hale derivatives. To compute each 2nd derivative in the most straightforward way, we must solve one forward problem F (m). Therefore, the Hessian matrix computed in this way requires the solution of O(N 2 ) forward problems. Pratt et al. (1998) show a more efficient method that reduces the number of required forward solutions from O(N 2 ) to O(N), which is still too costly in realistic large-size FWI problems. Moreover, the Hessian matrix consumes large amounts of memory. In a model with N parameters, a Hessian matrix stored in single precision requires 4N 2 bytes of memory. Because of the high cost of Newton s method for large problems, various methodologies have been proposed to approximate the Hessian matrix. The Gauss- Newton method, for example, ignores 2nd-order terms that account for nonlinearity in the function F (m). This approximation is valid only when the forward operator F is a linear or quasi-linear function of m. If the initial model in FWI is far from the true model and nonlinearity is significant, the Gauss-Newton method may not converge. Quasi-Newton methods do not compute the Hessian matrix explicitly, but instead iteratively update Hessian approximations. The BFGS method (Broyden, 1970; Fletcher, 1970; Goldfarb, 1970; Shanno, 1970) is an efficent way to update the Hessian matrix. However, although the BFGS method reduces the computation time required to approximate a Hessian matrix, it does not reduce the computation time required to use the Hessian matrix to update model parameters, nor does it decrease the amount of memory required to store Hessian approximations. In practical applications of quasi- Newton methods for FWI, we must reduce both computation time and memory consumption significantly. In this paper, we first review various Newton-like methods. We then pose FWI in a sparse model space and introduce a projected Hessian matrix in that space. We obtain the projected Hessian by using a projected BFGS method, a version of BFGS in the sparse model space. A projected Hessian matrix built in this way significantly reduces both computation time and memory consumption. Tests of our projected Hessian on the Marmousi II model suggest that quasi-newton methods like ours may be promising in practical applications of FWI. 2 NEWTON-LIKE METHODS We consider only iterative methods for solution of the nonlinear FWI optimization problem. In the i th iteration of such methods, we update the model m in the direction of a vector p i : m i+1 = m i + α ip i, (2) for some step length α i. Newton s method, the Gauss- Newton method, and quasi-newton methods differ in the way in which they compute and use the update vector p i. 2.1 Newton s method In Newton s method, the update vector p i is determined by first ignoring the higher-order (> 2) terms in a Taylor series approximation of the objective function E (m): E (m i + δm i) E (m i) + δm T i g i δmt i H iδm i, (3) where E (m i) denotes the objective function evaluated at m i, g i = g (m i), and H i = H (m i). For models m near the current model estimate m i, Equation 3 approximates the objective function with a quadratic function. We then minimize this approximated E (m) by solving a set of linear equations: with solution H iδm i = g i (4) δm i = H i g i. (5) Therefore, the update direction in Newton s method is simply p i = H i g i, (6) and the step length α i = 1. However, for models with large numbers of parameters, high costs for computing the Hessian matrix H i prevent the application of Newton s method. To analyze the cost, let us explicitly write each element of H i as H jk = 2 E m j m k, (7) where j and k are parameter indices in the model m, ranging from 1 to the total number N of model parameters. If, say, N = in a 2D FWI problem, the Hessian H i is a matrix. To compute this Hessian matrix in the most straightforward way, we must solve forward problems F (m) in each iteration of FWI. Even using the more efficient way in Pratt et al. (1998), we can only reduce this number to per iteration. For FWI, which normally requires multiple iterations, this cost is prohibitive. Furthermore, the memory required to store this Hessian matrix in single precision is 1 terabyte, or about 1/2 terabyte considering symmetry in the Hessian. 2.2 Gauss-Newton method Discarding the second term on the right-hand side of equation 7, we get an approximated Hessian H a, with elements H jk J T j J k, (8)

3 Projected Hessian 15 where J j and J k are the j th and k th columns in the Jacobian matrix { F(m) m 1, F(m) m 2,..., F(m) m N }, respectively. Using this approximation we obtain the Gauss-Newton method with update direction p i = H a g i. (9) Equation 8 shows that the cost of computing H a is equivalent to that of computing the gradient g i. Pratt et al. (1998) show an efficient approach for calculating g i, and only 2 forward data syntheses F (m) are necessary. However, the approximation H a in effect assumes that the function F (m i + δm i) is linear with respect to the difference δm i between the true model m and the current model m i. In practice this assumption is seldom realistic. 2.3 Quasi-Newton method A quasi-newton method iteratively approximates the Hessian matrix. The BFGS method (Broyden, 1970; Fletcher, 1970; Goldfarb, 1970; Shanno, 1970) is currently considered the most popular (Nocedal and Wright, 2000) quasi-newton update formula. The basic idea of BFGS is to update the Hessian using changes in both the model m i and the gradient g i, from one iteration to the next: H i+1 = H i + y iy T i Hiδmi (Hiδmi)T, (10) y T i δmi δm T i Hiδmi where y i = g i+1 g i, δm i = m i+1 m i. The BFGS method normally begins with an identity matrix H 0 = I in the first iteration, so that the first iteration of BFGS is similar to a gradientdescent method. Pseudo-code for implementing the BFGS method is as follows (Nocedal and Wright, 2000): given m 0, g 0 = E m 0, and H 0 = I for i = 0, 1, 2,..., until convergence do p i = H i g i search for α i, δm i = α ip i m i+1 = m i + δm i g i+1 = E (m i+1) y i = g i+1 g i H i+1 = H i + y i yt i H iδm i (H i δm i ) T y T i δm i δm T i H iδm i end for The BFGS method is an efficient way to compute Hessian approximations. Like the Gauss-Newton method, BFGS requires computation of only gradients g i, with the cost of only 2 forward modelings F (m). However, unlike the the Gauss-Newton method, BFGS does not simply ignore 2nd derivative terms in the Hessian. Instead, the BFGS method uses differences in models and gradients between iterations to construct a complete Hessian approximation. Unfortunately, the BFGS method does not reduce the O ( N 3) cost of solving the set of linear equations (equation 4) for the update vector p i, nor does it reduce the O ( N 2) memory required to store the Hessian approximation. 3 PROJECTED HESSIAN MATRIX The only way to reduce these costs is to reduce the number N of model parameters. For this purpose, we introduce a projected Hessian matrix in a sparse model space. 3.1 Sparse optimization In an approach similar to that used in subspace methods (Kennett et al., 1988; Oldenburg et al., 1993) and alternative parameterizations (Pratt et al., 1998, Appendix A), we construct a finely- and uniformly-sampled (dense) model m from a sparse model s that contains a much smaller number n << N of model parameters: m = Rs, (11) where R is an interpolation operator that interpolates model parameters from the sparse model s to the dense model m. FWI is then reformulated as a new sparse optimization problem (Pratt et al., 1998, Appendix A; Ma et al., 2010), in which we minimize a new nonlinear objective function E: R n R, min E (Rs). (12) s Rn In the optimization problem posed in equation 12, model parameters in m are not determined independently as in equation 1. Instead, the n sparse parameters in s constrain the other N n parameters in m. This constraint is imposed by the interpolation operator R. Therefore, the optimization problem in equation 12 is equivalent to a linearly constrained optimization problem in the dense model space m, where the number of constraints is N n. In the context of constrained optimization, the projected Hessian matrix is suggested by Gill et al. (1981) and several other authors. We rewrite the objective function in equation 3 as E (Rs i + Rδs i) E (Rs i) + δs T i R T g i δst i R T H irδs i, (13) where we refer to R T H ir as a projected Hessian matrix (Gill et al., 1981), and R T g i as a projected gradient. Here R T denotes the adjoint of the interpolation operator R. For models s near the current model estimate s i, equation 13 approximates the objective function with a quadratic function. We then minimize this approximated E (Rs) by solving a set of n linear equations: R T H irδs i = R T g i, (14)

4 R T H i R RHR 16 Y. Ma & D. Hale R T H i R = Hi Figure 2. Schematic for computing a projected Hessian H i. The projected Hessian reduces computation time and memory consumption. with a solution for the n unknowns ( δs i = R T H ir) R T g i. (15) Therefore, in Newton s method the update direction vector becomes ( p i = R T H ir) R T g i, (16) for a step length α i = 1. The projected Hessian R T H ir is a n n symmetric matrix. Figure 2 describes this projected Hessian matrix in a schematic fashion, where the tall and thin rectangle denotes the interpolation operator R, the short and fat rectangle denotes the adjoint operator R T, and the large and small squares denote the original and projected Hessian matrices, respectively. This schematic diagram illustrates how memory consumption is reduced from O ( N 2) to O ( n 2). Computation is likewise reduced from O ( N 3) to O ( n 3). 3.2 Projected BFGS method A projected Hessian matrix can be approximated iteratively with an approach similar to the classic BFGS method. Apply the interpolation operator R and its adjoint operator R T to both sides of equation 10 to obtain R T H i+1r = R T H ir + RT y i y T i R y T i δmi RT H iδm i ( R T H iδm i ) T δm T i Hiδmi. (17) Using the simple relationship δm i rewrite equation 17 as R T H i+1r = R T H ir + RT y i y T i R y T i Rδsi = = Rδs i, we can RT H irδs i ( R T H irδs i ) T δs T i RT H irδs i. (18) Now simplify equation 18 by defining H i R T H ir, g i R T g i, and ỹ i R T y i = g i+1 g i to obtain an update formula for the projected Hessian matrix: H i+1 = H i + ỹiỹt i ỹ T i δsi H iδs i ( Hiδs i ) T δs T i H iδs i. (19) We can further simplify equation 19 with δs i = α i p i and H iδs i = α i g i (Gill et al., 1981): H i+1 = H i + ỹiỹt i α iỹ T i p i + g i gt i p T i g. (20) i Equation 20 has the same structure as equation 10, so we refer to this formula for updating the projected Hessian as the projected BFGS method, a version of the classic BFGS method in the sparse model space s. Pseudo-code for implementing this projected BFGS method is as follows: given s 0, g 0 = R T g 0, and H 0 = I for i = 0, 1, 2,..., until convergence do p i = H i g i search for α i, δs i = α i p i m i+1 = m i + Rδs i g i+1 = E (m i+1), g i+1 = R T g i+1 ỹ i = g i+1 g i H i+1 = H i + (equation 20) end for 3.3 Wolfe conditions ỹiỹt i + g i gt α i ỹ T i i p i p T i g i The projected Hessian matrix H is symmetric positivedefinite (SPD) if ỹ T i δs i > 0 in equation 19. This SPD property is important for an optimization problem, because it guarantees that a quasi-newton method that uses this projected Hessian converges to (possibly local) minimum. This SPD property can be achieved if the step length α i satisfies the sufficient descent condition E (s i + α i p i ) E (α i) E (0) + c 1E (0) α i, (21) and the curvature condition E (α i) c 2E (0), (22) where E (0) = g T i p i and E (α i) = g T i+1 p i are directional derivatives. In equations 21 and 22, c 1 and c 2 are constants (0, 1). Suggested values for c 1 and c 2 are and 0.9 (Nocedal and Wright, 2000), respectively. Together, equations 21 and 22 are referred to as the Wolfe conditions (Nocedal and Wright, 2000). If the curvature condition in equation 22 is replaced by E (α i) c 2 E (0), (23) then the Wolfe conditions become the strong Wolfe conditions used in this study.

5 Projected Hessian 17 Figure 4. Structurally constrained sample selections. A total of 165 samples are chosen for our projected BFGS method. 4.1 Projected Hessian and its inverse Figure 3. The Marmousi II velocity model and an initial velocity model. 3.4 Cost of projected BFGS In each iteration of the projected BFGS method, we perform at least one forward calculation F (m) and compute at least one gradients g. Therefore, in the ideal case, the cost of each iteration of the projected BFGS method is that of three forward calculations F (m). Furthermore, if the number of parameters n in the sparse model space s is much smaller than the number N in the dense model m, the amount of memory required to store the projected Hessian is significantly reduced, as shown in Figure 2. 4 EXAMPLES We test our projected BFGS method on the Marmousi II model. Figure 3a shows the true model m, and Figure 3b shows the initial model m 0, a highly smoothed version of the true model. We use 11 shots uniformly distributed on the surface, and a 15Hz Ricker wavelet as the source for simulating wavefields. The source and receiver intervals are 0.76 km and km, respectively. In this example, the dense model space has N = parameters. Therefore, computation or storage of the true Hessian matrix for this example is infeasible. Our projected BFGS method poses the optimization problem in a sparse model space s. As described by Ma et al. (2011), we construct the sparse model s using a structurally constrained sample selection scheme. This selection method considers structures apparent in migrated images, and places samples along structural features. Figure 4 shows a total of 165 scattered sample locations. This set of scattered sample locations are representative, because (1) they lie between (not on) reflectors; (2) they are distributed along structural features; (3) there are more samples in structurally complex areas and fewer samples in simple areas. The scattered locations together with corresponding values at these locations comprise a sparse model space s that will be used in the projected BFGS method. We employ image-guided interpolation (IGI) (Hale, 2009) as the operator R and the adjoint of image-guided interpolation (Ma et al., 2010) as the operator R T. IGI interpolates values in the sparse model s to a uniformly sampled dense model m, and the interpolant makes good geological sense because it accounts for structures in the model. Figure 5a shows the initial projected Hessian H 0, an identity matrix. Figure 5b shows the updated projected Hessian H 1 after one iteration of the projected BFGS method. The update H 1 H 0 to the Hessian in this first iteration is shown separately in Figure 5c. As we can see, the projected BFGS method adds significant off-diagonal components to the initial Hessian H 0. Because our line search satisfies the strong Wolfe conditions, the inverse of the projected Hessian matrix exists. Figure 5d shows the inverse Hessian H 1.

6 18 Y. Ma & D. Hale (c) (d) Figure 5. The initial Hessian H 0, the Hessian matrix H 1 after the 1st iteration, the Hessian update H 1 H 0 in the 1st iteration (c), and the inverse Hessian H 1 (d). 4.2 Eigenvalues and eigenvectors Note that the last two terms of the right-hand side of equation 19 are two rank-one matrices, so each iteration of the projected BFGS method is a rank-two update to the previous Hessian approximation. Figures 6a, 6b, 6c and 6d show the eigenvalues of the projected Hessian H i in the 1st, 4th, 7th and 10th iterations, respectively. As suggested by the rank-two update, the projected BFGS method changes only two eigenvalues in each iteration: one, the smallest, and the other, the largest. Figures 7a and 7b show two eigenvectors of the projected Hessian H 1, corresponding to the largest and the smallest eigenvalues, respectively. The eigenvalues and eigenvectors of the projected Hessian have geometric meaning. Consider the objective function E (m) as a multidimensional parabolic bowl (Thacker, 1989). The eigenvector associated with the largest eigenvalue points in the direction of greatest curvature on the bowl. Updating the model m along the eigenvector direction shown in Figure 7a, we get the largest rate of change in the objective function E (m). This eigenvector indicates the component of the model m that can be best determined in FWI. Likewise, the eigenvector associated with the smallest eigenvalue indicates the direction of least curvature. Updating the model m along the eigenvector direction shown in Figure 7b yields the smallest rate of change of E (m). This eigenvector indicates the component of the model m that can be least well determined in FWI. 4.3 Projected quasi-newton FWI The projected BFGS method updates the model in each iteration, and therefore the projected BFGS method can be directly used in quasi-newton solutions to FWI. The key difference between quasi-newton FWI and

7 Projected Hessian Eigenvalues Eigenvalues Index Index 3.5 Eigenvalues Index (c) Eigenvalues Index (d) Figure 6. Eigenvalues of projected Hessian in 1st, 4th, 7th (c), and 10th (d) iterations. Figure 7. Eigenvectors of H 1 corresponding to the largest and the smallest eigenvalues. Geometrically, an eigenvector is a direction on the multidimensional surface of an objective function E (m), and the associated eigenvalue determines the curvature on the surface. If the model is updated in the eigenvector direction in, the rate of change of E (m) is largest. If the model is updated in the eigenvector direction in, the rate of change of E (m) is smallest.

8 20 Y. Ma & D. Hale Table 1. Update directions p i for the conjugate gradient (CG) method, the image-guided CG method, and our quasi- Newton method. Method Conjugate gradient Image-guided CG Quasi-Newton Update direction p 0 = g 0 β i β ( g i, g i ) = g T i (g i g i ) g T i g i p i = g i + β i p i p 0 = RR T g 0 β i = β ( RR T g i, RR T g i ) p i = RR T g i + β i p i H 0 = I p i = R H i R T g i other solutions, such as the conjugate gradient method and Newton s method, is the update direction. Table 1 compares the update directions used in the conjugate gradient method, the image-guided conjugate gradient method (Ma et al., 2011), and the quasi-newton method. As suggested in Ma et al. (2011), the image-guided conjugate gradient method uses a gather-scatter process in applying RR T, and thereby generates low wavenumbers not present in the simple conjugate gradient due to the lack of low frequencies in recorded data. The conjugate gradient in Figure 1a and the image-guided conjugate gradient in Figure 1b illustrate this difference. Figures 8a and 8c show inversion results estimated by the conjugate gradient (CG) method and the imageguided CG method, respectively, after two iterations. The corresponding estimated models in the 10th iteration are shown Figure 8b and 8d, respectively. We can see that both the conventional CG and the imageguided CG methods mainly update the shallow part of the model. This is due to the fact that both the conjugate gradient and the image-guided conjugate gradient have unbalanced amplitudes, high in the shallow part and low in the deeper part. In the case of quasi-newton FWI, the update direction shown in Figure 1c not only contains significant low wavenumbers, as does the image-guided conjugate gradient in Figure 1b, but the amplitudes are more balanced as well. In Figure 1c, the amplitudes are higher in the deeper part. This improvement is due to the use of H i the inverse projected Hessian matrix (see the last row in Table 1), which works as a filter applied to the gradient. Figures 8e and 8f show the models estimated in the 2nd and 10th iterations of the quasi-newton method, respectively. As we can see, quasi-newton FWI updates the deeper part of the model in early iterations. Figure 9 shows the convergence rates of the quasi- Newton method in this Marmousi II example with the data misfit function and the model misfit function (an L2 norm of the difference between the true model and the estimated model). Overall, the quasi-newton FWI shows faster convergence. 5 CONCLUSIONS A projected BFGS method iteratively approximates the Hessian matrix in FWI, thereby reducing both computation time and required memory. The projected BFGS method enables FWI to be performed using a quasi- Newton method. As demonstrated by the Marmousi II example, quasi-newton FWI converges in fewer iterations. Compared with the conjugate gradient method or the image-guided conjugate gradient method, the primary disadvantage of our projected BFGS method is the computational cost of a single iteration, which is relatively high because the line search must satisfy the Wolfe conditions. Therefore, a further investigation of the coefficients in the Wolfe conditions is necessary. 6 ACKNOWLEDGMENT This work is sponsored by a research agreement between ConocoPhillips and Colorado School of Mines (SST SRA). Yong Ma thanks Diane Witters for help in learning how to polish this manuscript. REFERENCES Broyden, C. G., 1970, The convergence of a class of double-rank minimization algorithms: IMA Journal of Applied Mathematics, 6, Fletcher, R., 1970, A new approach to variable metric algorithms: The Computer Journal, 13, Gill, P. E., W. Murray, and M. H. Wright, 1981, Practical optimization: Academic Press, London. Goldfarb, D., 1970, A family of variable-metric methods derived by variational means: Mathematics of Computation, 109, Hale, D., 2009, Image-guided blended neighbor interpolation of scattered data: SEG Technical Program Expanded Abstracts, 28, Kennett, B., M. Sambridge, and P. Williamson, 1988, Subspace methods for large inverse problems with multiple parameter classes: Geophysical Journal International, 94, Ma, Y., D. Hale, B. Gong, and Z. Meng, 2011, Imageguided full waveform inversion: CWP Report. Ma, Y., D. Hale, Z. Meng, and B. Gong, 2010, Full waveform inversion with image-guided gradient: SEG Technical Program Expanded Abstracts, 29,

9 Projected Hessian (c) (d) (e) (f) 21 Figure 8. Inverted velocity models for the conjugate gradient method (a,b), the image-guided conjugate gradient method (c,d), and the quasi-newton method (e,f). Left (a,c,e): the 2nd iteration; right (b,d,f): the 10th iteration. Nocedal, J., and S. J. Wright, 2000, Numerical Optimization: Springer. Oldenburg, D., P. McGillvray, and R. Ellis, 1993, Generalized subspace methods for large-scale inverse problems: Geophysical Journal International, 114, Pratt, R., C. Shin, and G. Hicks, 1998, Gauss-Newton and full newton methods in frequency-space seismic waveform inversion: Geophysical Journal International, 133, Pratt, R. G., 1999, Seismic waveform inversion in the frequency domain, part 1: Theory and verification in a physical scale model: Geophysics, 64, 888. Shanno, D. F., 1970, Conditioning of quasi-newton methods for function minimization: Mathematics of Computation, 111, Sheen, D. H., K. Tuncay, C. E. Baag, and P. J. Ortoleva, 2006, Time domain gauss-newton seismic waveform inversion in elastic media: Geophysical Journal International, 167, Symes, W. W., 2008, Migration velocity analysis and waveform inversion: Geophysical Prospecting, 56, Tarantola, A., 1984, Inversion of seismic-reflection data

10 22 Y. Ma & D. Hale Figure 9. Convergence of the conjugate gradient method, the image-guided conjugate gradient method, and the projected quasi-newton method: the data misfit function and the model misfit function. in the acoustic approximation: Geophysics, 49, , 2005, Inverse problem theory and methods for model parameter estimation: Society for Industrial and Applied Mathematics. Thacker, W. C., 1989, The role of the Hessian matrix in fitting models to measurements: Journal of Geophysical Research, 94, Virieux, J., and S. Operto, 2009, An overview of fullwaveform inversion in exploration geophysics: Geophysics, 74, WCC1 WCC26.

Comparison between least-squares reverse time migration and full-waveform inversion

Comparison between least-squares reverse time migration and full-waveform inversion Comparison between least-squares reverse time migration and full-waveform inversion Lei Yang, Daniel O. Trad and Wenyong Pan Summary The inverse problem in exploration geophysics usually consists of two

More information

SUMMARY REVIEW OF THE FREQUENCY DOMAIN L2 FWI-HESSIAN

SUMMARY REVIEW OF THE FREQUENCY DOMAIN L2 FWI-HESSIAN Efficient stochastic Hessian estimation for full waveform inversion Lucas A. Willemsen, Alison E. Malcolm and Russell J. Hewett, Massachusetts Institute of Technology SUMMARY In this abstract we present

More information

Uncertainty quantification for Wavefield Reconstruction Inversion

Uncertainty quantification for Wavefield Reconstruction Inversion Uncertainty quantification for Wavefield Reconstruction Inversion Zhilong Fang *, Chia Ying Lee, Curt Da Silva *, Felix J. Herrmann *, and Rachel Kuske * Seismic Laboratory for Imaging and Modeling (SLIM),

More information

Approximate- vs. full-hessian in FWI: 1D analytical and numerical experiments

Approximate- vs. full-hessian in FWI: 1D analytical and numerical experiments Approximate- vs. full-hessian in FWI: 1D analytical and numerical experiments Raul Cova and Kris Innanen ABSTRACT Feasibility of using Full Waveform Inversion (FWI) to build velocity models has been increasing

More information

Source estimation for frequency-domain FWI with robust penalties

Source estimation for frequency-domain FWI with robust penalties Source estimation for frequency-domain FWI with robust penalties Aleksandr Y. Aravkin, Tristan van Leeuwen, Henri Calandra, and Felix J. Herrmann Dept. of Earth and Ocean sciences University of British

More information

Optimization Methods

Optimization Methods Optimization Methods Decision making Examples: determining which ingredients and in what quantities to add to a mixture being made so that it will meet specifications on its composition allocating available

More information

Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno February 6, / 25 (BFG. Limited memory BFGS (L-BFGS)

Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno February 6, / 25 (BFG. Limited memory BFGS (L-BFGS) Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno (BFGS) Limited memory BFGS (L-BFGS) February 6, 2014 Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb

More information

2. Quasi-Newton methods

2. Quasi-Newton methods L. Vandenberghe EE236C (Spring 2016) 2. Quasi-Newton methods variable metric methods quasi-newton methods BFGS update limited-memory quasi-newton methods 2-1 Newton method for unconstrained minimization

More information

Improving the Convergence of Back-Propogation Learning with Second Order Methods

Improving the Convergence of Back-Propogation Learning with Second Order Methods the of Back-Propogation Learning with Second Order Methods Sue Becker and Yann le Cun, Sept 1988 Kasey Bray, October 2017 Table of Contents 1 with Back-Propagation 2 the of BP 3 A Computationally Feasible

More information

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09

Numerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09 Numerical Optimization 1 Working Horse in Computer Vision Variational Methods Shape Analysis Machine Learning Markov Random Fields Geometry Common denominator: optimization problems 2 Overview of Methods

More information

Higher-Order Methods

Higher-Order Methods Higher-Order Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. PCMI, July 2016 Stephen Wright (UW-Madison) Higher-Order Methods PCMI, July 2016 1 / 25 Smooth

More information

FWI with Compressive Updates Aleksandr Aravkin, Felix Herrmann, Tristan van Leeuwen, Xiang Li, James Burke

FWI with Compressive Updates Aleksandr Aravkin, Felix Herrmann, Tristan van Leeuwen, Xiang Li, James Burke Consortium 2010 FWI with Compressive Updates Aleksandr Aravkin, Felix Herrmann, Tristan van Leeuwen, Xiang Li, James Burke SLIM University of British Columbia Full Waveform Inversion The Full Waveform

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428

More information

Wavefield Reconstruction Inversion (WRI) a new take on wave-equation based inversion Felix J. Herrmann

Wavefield Reconstruction Inversion (WRI) a new take on wave-equation based inversion Felix J. Herrmann Wavefield Reconstruction Inversion (WRI) a new take on wave-equation based inversion Felix J. Herrmann SLIM University of British Columbia van Leeuwen, T and Herrmann, F J (2013). Mitigating local minima

More information

Full-Waveform Inversion with Gauss- Newton-Krylov Method

Full-Waveform Inversion with Gauss- Newton-Krylov Method Full-Waveform Inversion with Gauss- Newton-Krylov Method Yogi A. Erlangga and Felix J. Herrmann {yerlangga,fherrmann}@eos.ubc.ca Seismic Laboratory for Imaging and Modeling The University of British Columbia

More information

LBFGS. John Langford, Large Scale Machine Learning Class, February 5. (post presentation version)

LBFGS. John Langford, Large Scale Machine Learning Class, February 5. (post presentation version) LBFGS John Langford, Large Scale Machine Learning Class, February 5 (post presentation version) We are still doing Linear Learning Features: a vector x R n Label: y R Goal: Learn w R n such that ŷ w (x)

More information

Quasi-Newton Methods

Quasi-Newton Methods Newton s Method Pros and Cons Quasi-Newton Methods MA 348 Kurt Bryan Newton s method has some very nice properties: It s extremely fast, at least once it gets near the minimum, and with the simple modifications

More information

5 Quasi-Newton Methods

5 Quasi-Newton Methods Unconstrained Convex Optimization 26 5 Quasi-Newton Methods If the Hessian is unavailable... Notation: H = Hessian matrix. B is the approximation of H. C is the approximation of H 1. Problem: Solve min

More information

Methods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent

Methods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent Nonlinear Optimization Steepest Descent and Niclas Börlin Department of Computing Science Umeå University niclas.borlin@cs.umu.se A disadvantage with the Newton method is that the Hessian has to be derived

More information

Convex Optimization CMU-10725

Convex Optimization CMU-10725 Convex Optimization CMU-10725 Quasi Newton Methods Barnabás Póczos & Ryan Tibshirani Quasi Newton Methods 2 Outline Modified Newton Method Rank one correction of the inverse Rank two correction of the

More information

Migration with Implicit Solvers for the Time-harmonic Helmholtz

Migration with Implicit Solvers for the Time-harmonic Helmholtz Migration with Implicit Solvers for the Time-harmonic Helmholtz Yogi A. Erlangga, Felix J. Herrmann Seismic Laboratory for Imaging and Modeling, The University of British Columbia {yerlangga,fherrmann}@eos.ubc.ca

More information

Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

More information

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Multidimensional Unconstrained Optimization Suppose we have a function f() of more than one

More information

ORIE 6326: Convex Optimization. Quasi-Newton Methods

ORIE 6326: Convex Optimization. Quasi-Newton Methods ORIE 6326: Convex Optimization Quasi-Newton Methods Professor Udell Operations Research and Information Engineering Cornell April 10, 2017 Slides on steepest descent and analysis of Newton s method adapted

More information

The truncated Newton method for Full Waveform Inversion

The truncated Newton method for Full Waveform Inversion The truncated Newton method for Full Waveform Inversion Ludovic Métivier, Romain Brossier, Jean Virieux, Stéphane Operto To cite this version: Ludovic Métivier, Romain Brossier, Jean Virieux, Stéphane

More information

Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning

Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning Jacob Rafati http://rafati.net jrafatiheravi@ucmerced.edu Ph.D. Candidate, Electrical Engineering and Computer Science University

More information

Waveform inversion for attenuation estimation in anisotropic media Tong Bai & Ilya Tsvankin Center for Wave Phenomena, Colorado School of Mines

Waveform inversion for attenuation estimation in anisotropic media Tong Bai & Ilya Tsvankin Center for Wave Phenomena, Colorado School of Mines Waveform inversion for attenuation estimation in anisotropic media Tong Bai & Ilya Tsvankin Center for Wave Phenomena, Colorado School of Mines SUMMARY Robust estimation of attenuation coefficients remains

More information

NonlinearOptimization

NonlinearOptimization 1/35 NonlinearOptimization Pavel Kordík Department of Computer Systems Faculty of Information Technology Czech Technical University in Prague Jiří Kašpar, Pavel Tvrdík, 2011 Unconstrained nonlinear optimization,

More information

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23 Optimization: Nonlinear Optimization without Constraints Nonlinear Optimization without Constraints 1 / 23 Nonlinear optimization without constraints Unconstrained minimization min x f(x) where f(x) is

More information

Algorithms for Constrained Optimization

Algorithms for Constrained Optimization 1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic

More information

MS&E 318 (CME 338) Large-Scale Numerical Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization 1 Origins Instructor: Michael Saunders Spring 2015 Notes 9: Augmented Lagrangian Methods

More information

Elastic least-squares migration with two-way wave equation forward and adjoint operators

Elastic least-squares migration with two-way wave equation forward and adjoint operators Elastic least-squares migration with two-way wave equation forward and adjoint operators Ke Chen and Mauricio D. Sacchi, Department of Physics, University of Alberta Summary Time domain elastic least-squares

More information

Full waveform inversion and the inverse Hessian

Full waveform inversion and the inverse Hessian Full waveform inversion and the inverse Hessian Gary Margrave, Matt Yedlin and Kris Innanen ABSTRACT FWI and the inverse Hessian Full waveform inversion involves defining an objective function, and then

More information

W011 Full Waveform Inversion for Detailed Velocity Model Building

W011 Full Waveform Inversion for Detailed Velocity Model Building W011 Full Waveform Inversion for Detailed Velocity Model Building S. Kapoor* (WesternGeco, LLC), D. Vigh (WesternGeco), H. Li (WesternGeco) & D. Derharoutian (WesternGeco) SUMMARY An accurate earth model

More information

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Newton s Method. Ryan Tibshirani Convex Optimization /36-725 Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x

More information

Optimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng

Optimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng Optimization 2 CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Optimization 2 1 / 38

More information

Nonlinear Optimization Methods for Machine Learning

Nonlinear Optimization Methods for Machine Learning Nonlinear Optimization Methods for Machine Learning Jorge Nocedal Northwestern University University of California, Davis, Sept 2018 1 Introduction We don t really know, do we? a) Deep neural networks

More information

Full-waveform inversion application in different geological settings Denes Vigh*, Jerry Kapoor and Hongyan Li, WesternGeco

Full-waveform inversion application in different geological settings Denes Vigh*, Jerry Kapoor and Hongyan Li, WesternGeco Full-waveform inversion application in different geological settings Denes Vigh*, Jerry Kapoor and Hongyan Li, WesternGeco Summary After the synthetic data inversion examples, real 3D data sets have been

More information

2D Laplace-Domain Waveform Inversion of Field Data Using a Power Objective Function

2D Laplace-Domain Waveform Inversion of Field Data Using a Power Objective Function Pure Appl. Geophys. Ó 213 Springer Basel DOI 1.17/s24-13-651-4 Pure and Applied Geophysics 2D Laplace-Domain Waveform Inversion of Field Data Using a Power Objective Function EUNJIN PARK, 1 WANSOO HA,

More information

A 27-point scheme for a 3D frequency-domain scalar wave equation based on an average-derivative method

A 27-point scheme for a 3D frequency-domain scalar wave equation based on an average-derivative method Geophysical Prospecting 04 6 58 77 doi: 0./365-478.090 A 7-point scheme for a 3D frequency-domain scalar wave equation based on an average-derivative method Jing-Bo Chen Key Laboratory of Petroleum Resources

More information

Youzuo Lin and Lianjie Huang

Youzuo Lin and Lianjie Huang PROCEEDINGS, Thirty-Ninth Workshop on Geothermal Reservoir Engineering Stanford University, Stanford, California, February 24-26, 2014 SGP-TR-202 Building Subsurface Velocity Models with Sharp Interfaces

More information

Chapter 4. Unconstrained optimization

Chapter 4. Unconstrained optimization Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file

More information

SUMMARY. (Sun and Symes, 2012; Biondi and Almomin, 2012; Almomin and Biondi, 2012).

SUMMARY. (Sun and Symes, 2012; Biondi and Almomin, 2012; Almomin and Biondi, 2012). Inversion Velocity Analysis via Differential Semblance Optimization in the Depth-oriented Extension Yujin Liu, William W. Symes, Yin Huang,and Zhenchun Li, China University of Petroleum (Huadong), Rice

More information

Tu SRS2 07 A Full-waveform Inversion Case Study from Offshore Gabon

Tu SRS2 07 A Full-waveform Inversion Case Study from Offshore Gabon Tu SRS2 07 A Full-waveform Inversion Case Study from Offshore Gabon A. Privitera (CGG), A. Ratcliffe* (CGG) & N. Kotova (CGG) SUMMARY We applied full waveform inversion to a seismic dataset from offshore

More information

Optimization schemes for Full Waveform Inversion: the preconditioned truncated Newton method

Optimization schemes for Full Waveform Inversion: the preconditioned truncated Newton method Optimization schemes for Full Waveform Inversion: the preconditioned truncated Newton method Ludovic Métivier, Romain Brossier, Jean Virieux, Stéphane Operto To cite this version: Ludovic Métivier, Romain

More information

Seismic Modeling, Migration and Velocity Inversion

Seismic Modeling, Migration and Velocity Inversion Seismic Modeling, Migration and Velocity Inversion Inverse Scattering Bee Bednar Panorama Technologies, Inc. 14811 St Marys Lane, Suite 150 Houston TX 77079 May 30, 2014 Bee Bednar (Panorama Technologies)

More information

Wenyong Pan and Lianjie Huang. Los Alamos National Laboratory, Geophysics Group, MS D452, Los Alamos, NM 87545, USA

Wenyong Pan and Lianjie Huang. Los Alamos National Laboratory, Geophysics Group, MS D452, Los Alamos, NM 87545, USA PROCEEDINGS, 44th Workshop on Geothermal Reservoir Engineering Stanford University, Stanford, California, February 11-13, 019 SGP-TR-14 Adaptive Viscoelastic-Waveform Inversion Using the Local Wavelet

More information

Full Waveform Inversion (FWI) with wave-equation migration. Gary Margrave Rob Ferguson Chad Hogan Banff, 3 Dec. 2010

Full Waveform Inversion (FWI) with wave-equation migration. Gary Margrave Rob Ferguson Chad Hogan Banff, 3 Dec. 2010 Full Waveform Inversion (FWI) with wave-equation migration (WEM) and well control Gary Margrave Rob Ferguson Chad Hogan Banff, 3 Dec. 2010 Outline The FWI cycle The fundamental theorem of FWI Understanding

More information

Unconstrained optimization

Unconstrained optimization Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout

More information

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality

More information

Conjugate Directions for Stochastic Gradient Descent

Conjugate Directions for Stochastic Gradient Descent Conjugate Directions for Stochastic Gradient Descent Nicol N Schraudolph Thore Graepel Institute of Computational Science ETH Zürich, Switzerland {schraudo,graepel}@infethzch Abstract The method of conjugate

More information

GEOPHYSICAL INVERSE THEORY AND REGULARIZATION PROBLEMS

GEOPHYSICAL INVERSE THEORY AND REGULARIZATION PROBLEMS Methods in Geochemistry and Geophysics, 36 GEOPHYSICAL INVERSE THEORY AND REGULARIZATION PROBLEMS Michael S. ZHDANOV University of Utah Salt Lake City UTAH, U.S.A. 2OO2 ELSEVIER Amsterdam - Boston - London

More information

Statistics 580 Optimization Methods

Statistics 580 Optimization Methods Statistics 580 Optimization Methods Introduction Let fx be a given real-valued function on R p. The general optimization problem is to find an x ɛ R p at which fx attain a maximum or a minimum. It is of

More information

Quasi-Newton Methods. Javier Peña Convex Optimization /36-725

Quasi-Newton Methods. Javier Peña Convex Optimization /36-725 Quasi-Newton Methods Javier Peña Convex Optimization 10-725/36-725 Last time: primal-dual interior-point methods Consider the problem min x subject to f(x) Ax = b h(x) 0 Assume f, h 1,..., h m are convex

More information

Simultaneous estimation of wavefields & medium parameters

Simultaneous estimation of wavefields & medium parameters Simultaneous estimation of wavefields & medium parameters reduced-space versus full-space waveform inversion Bas Peters, Felix J. Herrmann Workshop W- 2, The limit of FWI in subsurface parameter recovery.

More information

Quasi-Newton methods for minimization

Quasi-Newton methods for minimization Quasi-Newton methods for minimization Lectures for PHD course on Numerical optimization Enrico Bertolazzi DIMS Universitá di Trento November 21 December 14, 2011 Quasi-Newton methods for minimization 1

More information

SUMMARY INTRODUCTION. f ad j (t) = 2 Es,r. The kernel

SUMMARY INTRODUCTION. f ad j (t) = 2 Es,r. The kernel The failure mode of correlation focusing for model velocity estimation Hyoungsu Baek 1(*), Henri Calandra 2, and Laurent Demanet 1 1 Dept. of Mathematics and Earth Resources Lab, Massachusetts Institute

More information

Quasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization

Quasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization Quasi-Newton Methods Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization 10-725 Last time: primal-dual interior-point methods Given the problem min x f(x) subject to h(x)

More information

Lecture 14: October 17

Lecture 14: October 17 1-725/36-725: Convex Optimization Fall 218 Lecture 14: October 17 Lecturer: Lecturer: Ryan Tibshirani Scribes: Pengsheng Guo, Xian Zhou Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

Gradient Descent. Dr. Xiaowei Huang

Gradient Descent. Dr. Xiaowei Huang Gradient Descent Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Three machine learning algorithms: decision tree learning k-nn linear regression only optimization objectives are discussed,

More information

Towards full waveform inversion: A torturous path

Towards full waveform inversion: A torturous path FWI Towards full waveform inversion: A torturous path J. Helen Isaac and Gary F. Margrave ABSTRACT Full waveform inversion (FWI) can be viewed as an iterative cycle involving forward modelling, pre-stack

More information

Optimization II: Unconstrained Multivariable

Optimization II: Unconstrained Multivariable Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Optimization II: Unconstrained Multivariable 1

More information

P016 Toward Gauss-Newton and Exact Newton Optimization for Full Waveform Inversion

P016 Toward Gauss-Newton and Exact Newton Optimization for Full Waveform Inversion P016 Toward Gauss-Newton and Exact Newton Optiization for Full Wavefor Inversion L. Métivier* ISTerre, R. Brossier ISTerre, J. Virieux ISTerre & S. Operto Géoazur SUMMARY Full Wavefor Inversion FWI applications

More information

Source function estimation in extended full waveform inversion

Source function estimation in extended full waveform inversion Source function estimation in extended full waveform inversion Lei Fu The Rice Inversion Project (TRIP) May 6, 2014 Lei Fu (TRIP) Source estimation in EFWI May 6, 2014 1 Overview Objective Recover Earth

More information

Optimization and Root Finding. Kurt Hornik

Optimization and Root Finding. Kurt Hornik Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding

More information

MATH 4211/6211 Optimization Quasi-Newton Method

MATH 4211/6211 Optimization Quasi-Newton Method MATH 4211/6211 Optimization Quasi-Newton Method Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Quasi-Newton Method Motivation:

More information

Elastic least-squares reverse time migration

Elastic least-squares reverse time migration CWP-865 Elastic least-squares reverse time migration Yuting Duan, Paul Sava, and Antoine Guitton Center for Wave Phenomena, Colorado School of Mines ABSTRACT Least-squares migration (LSM) can produce images

More information

Preconditioned conjugate gradient algorithms with column scaling

Preconditioned conjugate gradient algorithms with column scaling Proceedings of the 47th IEEE Conference on Decision and Control Cancun, Mexico, Dec. 9-11, 28 Preconditioned conjugate gradient algorithms with column scaling R. Pytla Institute of Automatic Control and

More information

arxiv: v1 [math.na] 1 Apr 2015

arxiv: v1 [math.na] 1 Apr 2015 Nonlinear seismic imaging via reduced order model backprojection Alexander V. Mamonov, University of Houston; Vladimir Druskin and Mikhail Zaslavsky, Schlumberger arxiv:1504.00094v1 [math.na] 1 Apr 2015

More information

Optimization Methods

Optimization Methods Optimization Methods Categorization of Optimization Problems Continuous Optimization Discrete Optimization Combinatorial Optimization Variational Optimization Common Optimization Concepts in Computer Vision

More information

Introduction to Seismic Imaging

Introduction to Seismic Imaging Introduction to Seismic Imaging Alison Malcolm Department of Earth, Atmospheric and Planetary Sciences MIT August 20, 2010 Outline Introduction Why we image the Earth How data are collected Imaging vs

More information

ON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS

ON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS ON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS Anders FORSGREN Tove ODLAND Technical Report TRITA-MAT-203-OS-03 Department of Mathematics KTH Royal

More information

Seismic imaging and optimal transport

Seismic imaging and optimal transport Seismic imaging and optimal transport Bjorn Engquist In collaboration with Brittany Froese, Sergey Fomel and Yunan Yang Brenier60, Calculus of Variations and Optimal Transportation, Paris, January 10-13,

More information

Attenuation compensation in least-squares reverse time migration using the visco-acoustic wave equation

Attenuation compensation in least-squares reverse time migration using the visco-acoustic wave equation Attenuation compensation in least-squares reverse time migration using the visco-acoustic wave equation Gaurav Dutta, Kai Lu, Xin Wang and Gerard T. Schuster, King Abdullah University of Science and Technology

More information

Lecture 7 Unconstrained nonlinear programming

Lecture 7 Unconstrained nonlinear programming Lecture 7 Unconstrained nonlinear programming Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University,

More information

Full waveform inversion with wave equation migration and well control

Full waveform inversion with wave equation migration and well control FWI by WEM Full waveform inversion with wave equation migration and well control Gary F. Margrave, Robert J. Ferguson and Chad M. Hogan ABSTRACT We examine the key concepts in full waveform inversion (FWI)

More information

THE RICE INVERSION PROJECT

THE RICE INVERSION PROJECT THE RICE INVERSION PROJECT Mario Bencomo, Lei Fu, Jie Hou, Guanghui Huang, Rami Nammour, and William Symes Annual Report 6 Copyright 5-6 by Rice University i TRIP6 TABLE OF CONTENTS William W. Symes, The

More information

Attenuation compensation in viscoacoustic reserve-time migration Jianyong Bai*, Guoquan Chen, David Yingst, and Jacques Leveille, ION Geophysical

Attenuation compensation in viscoacoustic reserve-time migration Jianyong Bai*, Guoquan Chen, David Yingst, and Jacques Leveille, ION Geophysical Attenuation compensation in viscoacoustic reserve-time migration Jianyong Bai*, Guoquan Chen, David Yingst, and Jacques Leveille, ION Geophysical Summary Seismic waves are attenuated during propagation.

More information

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method. Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization

More information

Optimization II: Unconstrained Multivariable

Optimization II: Unconstrained Multivariable Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Optimization II: Unconstrained

More information

Geophysical Journal International

Geophysical Journal International Geophysical Journal International Geophys. J. Int. (2012) 188, 1221 1242 doi: 10.1111/j.1365-246X.2011.05314.x Full waveform inversion strategy for density in the frequency domain Woodon Jeong, 1 Ho-Yong

More information

Numerical Optimization of Partial Differential Equations

Numerical Optimization of Partial Differential Equations Numerical Optimization of Partial Differential Equations Part I: basic optimization concepts in R n Bartosz Protas Department of Mathematics & Statistics McMaster University, Hamilton, Ontario, Canada

More information

8 Numerical methods for unconstrained problems

8 Numerical methods for unconstrained problems 8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields

More information

Wave-equation Hessian by phase encoding

Wave-equation Hessian by phase encoding Wave-equation Hessian by phase encoding Yaxun Tang ABSTRACT I present a method for computing wave-equation Hessian operators, also known as resolution functions or point-spread functions, under the Born

More information

Exploring the energy landscape

Exploring the energy landscape Exploring the energy landscape ChE210D Today's lecture: what are general features of the potential energy surface and how can we locate and characterize minima on it Derivatives of the potential energy

More information

Truncated Conjugate Gradient Method for History Matching in Reservoir Simulation

Truncated Conjugate Gradient Method for History Matching in Reservoir Simulation Trabalho apresentado no CNAC, Gramado - RS, 2016. Proceeding Series of the Brazilian Society of Computational and Applied athematics Truncated Conjugate Gradient ethod for History atching in Reservoir

More information

Registration-guided least-squares waveform inversion

Registration-guided least-squares waveform inversion Registration-guided least-squares waveform inversion Hyoungsu Baek 1, Henri Calandra, Laurent Demanet 1 1 MIT Mathematics department, TOTAL S.A. January 15 013 Abstract Full waveform inversion with frequency

More information

Numerical Optimization: Basic Concepts and Algorithms

Numerical Optimization: Basic Concepts and Algorithms May 27th 2015 Numerical Optimization: Basic Concepts and Algorithms R. Duvigneau R. Duvigneau - Numerical Optimization: Basic Concepts and Algorithms 1 Outline Some basic concepts in optimization Some

More information

Anisotropic Seismic Imaging and Inversion for Subsurface Characterization at the Blue Mountain Geothermal Field in Nevada

Anisotropic Seismic Imaging and Inversion for Subsurface Characterization at the Blue Mountain Geothermal Field in Nevada PROCEEDINGS, 43rd Workshop on Geothermal Reservoir Engineering Stanford University, Stanford, California, February 12-14, 2018 SGP-TR-213 Anisotropic Seismic Imaging and Inversion for Subsurface Characterization

More information

23855 Rock Physics Constraints on Seismic Inversion

23855 Rock Physics Constraints on Seismic Inversion 23855 Rock Physics Constraints on Seismic Inversion M. Sams* (Ikon Science Ltd) & D. Saussus (Ikon Science) SUMMARY Seismic data are bandlimited, offset limited and noisy. Consequently interpretation of

More information

ALGORITHM XXX: SC-SR1: MATLAB SOFTWARE FOR SOLVING SHAPE-CHANGING L-SR1 TRUST-REGION SUBPROBLEMS

ALGORITHM XXX: SC-SR1: MATLAB SOFTWARE FOR SOLVING SHAPE-CHANGING L-SR1 TRUST-REGION SUBPROBLEMS ALGORITHM XXX: SC-SR1: MATLAB SOFTWARE FOR SOLVING SHAPE-CHANGING L-SR1 TRUST-REGION SUBPROBLEMS JOHANNES BRUST, OLEG BURDAKOV, JENNIFER B. ERWAY, ROUMMEL F. MARCIA, AND YA-XIANG YUAN Abstract. We present

More information

ECS550NFB Introduction to Numerical Methods using Matlab Day 2

ECS550NFB Introduction to Numerical Methods using Matlab Day 2 ECS550NFB Introduction to Numerical Methods using Matlab Day 2 Lukas Laffers lukas.laffers@umb.sk Department of Mathematics, University of Matej Bel June 9, 2015 Today Root-finding: find x that solves

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

MS&E 318 (CME 338) Large-Scale Numerical Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization Instructor: Michael Saunders Spring 2015 Notes 11: NPSOL and SNOPT SQP Methods 1 Overview

More information

Mobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti

Mobile Robotics 1. A Compact Course on Linear Algebra. Giorgio Grisetti Mobile Robotics 1 A Compact Course on Linear Algebra Giorgio Grisetti SA-1 Vectors Arrays of numbers They represent a point in a n dimensional space 2 Vectors: Scalar Product Scalar-Vector Product Changes

More information

On Lagrange multipliers of trust-region subproblems

On Lagrange multipliers of trust-region subproblems On Lagrange multipliers of trust-region subproblems Ladislav Lukšan, Ctirad Matonoha, Jan Vlček Institute of Computer Science AS CR, Prague Programy a algoritmy numerické matematiky 14 1.- 6. června 2008

More information

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α

More information

P276 Multiparameter Full-Waveform Inversion for Velocity and Attenuation Refining the Imaging of a Sedimentary Basin

P276 Multiparameter Full-Waveform Inversion for Velocity and Attenuation Refining the Imaging of a Sedimentary Basin P276 Multiparameter Full-Waveform Inversion for Velocity and Attenuation Refining the Imaging of a Sedimentary Basin M. Malinowski* (Institute of Geophysics PAS), A. Ribodetti (Geosciences Azur CNRS UMR)

More information

REPORTS IN INFORMATICS

REPORTS IN INFORMATICS REPORTS IN INFORMATICS ISSN 0333-3590 A class of Methods Combining L-BFGS and Truncated Newton Lennart Frimannslund Trond Steihaug REPORT NO 319 April 2006 Department of Informatics UNIVERSITY OF BERGEN

More information

Overview Avoiding Cycle Skipping: Model Extension MSWI: Space-time Extension Numerical Examples

Overview Avoiding Cycle Skipping: Model Extension MSWI: Space-time Extension Numerical Examples Overview Avoiding Cycle Skipping: Model Extension MSWI: Space-time Extension Numerical Examples Guanghui Huang Education University of Chinese Academy of Sciences, Beijing, China Ph.D. in Computational

More information