Adaptive multiple subtraction using regularized nonstationary regression

Similar documents
Regularizing seismic inverse problems by model reparameterization using plane-wave construction

Short Note. Plane wave prediction in 3-D. Sergey Fomel 1 INTRODUCTION

Adaptive linear prediction filtering for random noise attenuation Mauricio D. Sacchi* and Mostafa Naghizadeh, University of Alberta

Stanford Exploration Project, Report 111, June 9, 2002, pages INTRODUCTION FROM THE FILTERING TO THE SUBTRACTION OF NOISE

SUMMARY ANGLE DECOMPOSITION INTRODUCTION. A conventional cross-correlation imaging condition for wave-equation migration is (Claerbout, 1985)

SUMMARY INTRODUCTION LOCALIZED PHASE ESTIMATION

Multiple realizations: Model variance and data uncertainty

On non-stationary convolution and inverse convolution

Bureau of Economic Geology, The University of Texas at Austin, Austin, TX. Research Associate, Geophysics from 11/2004

Effect of velocity uncertainty on amplitude information

Deconvolution imaging condition for reverse-time migration

Trace balancing with PEF plane annihilators

Stanford Exploration Project, Report 115, May 22, 2004, pages

Implicit 3-D depth migration by wavefield extrapolation with helical boundary conditions

P S-wave polarity reversal in angle domain common-image gathers

Multiple realizations using standard inversion techniques a

Stanford Exploration Project, Report 105, September 5, 2000, pages 41 53

Fast wavefield extrapolation by phase-shift in the nonuniform Gabor domain

Incorporating geologic information into reflection tomography

Fast wavefield extrapolation by phase-shift in the nonuniform Gabor domain

Stepwise Multiple Elimination using Linear Transforms: An alternative approach to SRME for stacking multiple-free near offsets in the Barents Sea

Structure-constrained relative acoustic impedance using stratigraphic coordinates a

Summary. Introduction

Velocity analysis using AB semblance a

Calculation of the sun s acoustic impulse response by multidimensional

Data examples of logarithm Fourier domain bidirectional deconvolution

A Padé approximation to the scalar wavefield extrapolator for inhomogeneous media

Direct nonlinear traveltime inversion in layered VTI media Paul J. Fowler*, Alexander Jackson, Joseph Gaffney, and David Boreham, WesternGeco

Pluto 1.5 2D ELASTIC MODEL FOR WAVEFIELD INVESTIGATIONS OF SUBSALT OBJECTIVES, DEEP WATER GULF OF MEXICO*

Smoothing along geologic dip

Iterative Methods for Smooth Objective Functions

One-step extrapolation method for reverse time migration

Seismic tomography with co-located soft data

Investigating fault shadows in a normally faulted geology

Controlled source interferometry with noisy data Jürg Hunziker, Joost van der Neut, Evert Slob and Kees Wapenaar, Delft University of Technology

Morse, P. and H. Feshbach, 1953, Methods of Theoretical Physics: Cambridge University

Comparison between least-squares reverse time migration and full-waveform inversion

Shot-profile migration of multiple reflections

The Deconvolution of Multicomponent Trace Vectors

Interpolation of nonstationary seismic records using a fast non-redundant S-transform

Time-migration velocity analysis by velocity continuation

Edinburgh Research Explorer

Residual statics analysis by LSQR

Compensating for attenuation by inverse Q filtering. Carlos A. Montaña Dr. Gary F. Margrave

Effect of Noise in Blending and Deblending Guus Berkhout* and Gerrit Blacquière, Delft University of Technology

COMPLEX TRACE ANALYSIS OF SEISMIC SIGNAL BY HILBERT TRANSFORM

for multicomponent seismic data Wave-equation angle-domain imaging

A Southern North Sea Multi-Survey presdm using Hybrid Gridded Tomography

Estimating a pseudounitary operator for velocity-stack inversion

Robust one-step (deconvolution + integration) seismic inversion in the frequency domain Ivan Priezzhev* and Aaron Scollard, Schlumberger

23855 Rock Physics Constraints on Seismic Inversion

Exploring three-dimensional implicit wavefield extrapolation with the helix transform a

Integration of seismic and fluid-flow data: a two-way road linked by rock physics

Elastic impedance inversion from robust regression method

Seismic attributes of time-vs. depth-migrated data using self-adaptive window

Multiple horizons mapping: A better approach for maximizing the value of seismic data

Improved image aids interpretation: A case history

H005 Pre-salt Depth Imaging of the Deepwater Santos Basin, Brazil

Nonstationary filters, pseudodifferential operators, and their inverses

Stanford Exploration Project, Report SERGEY, November 9, 2000, pages 683??

Rough sea estimation for phase-shift de-ghosting Sergio Grion*, Rob Telling and Seb Holland, Dolphin Geophysical

SEG/New Orleans 2006 Annual Meeting. Non-orthogonal Riemannian wavefield extrapolation Jeff Shragge, Stanford University

Local discontinuity measures for 3-D seismic data

Examples of prestack depth migration in TI media

Elastic least-squares reverse time migration

AVAZ inversion for fracture orientation and intensity: a physical modeling study

WE SRS2 11 ADAPTIVE LEAST-SQUARES RTM FOR SUBSALT IMAGING

Magnetic Case Study: Raglan Mine Laura Davis May 24, 2006

Downloaded 11/08/14 to Redistribution subject to SEG license or copyright; see Terms of Use at

Time-frequency analysis of seismic data using synchrosqueezing wavelet transform a

Compensating visco-acoustic effects in anisotropic resverse-time migration Sang Suh, Kwangjin Yoon, James Cai, and Bin Wang, TGS

Compensating for attenuation by inverse Q filtering

V003 How Reliable Is Statistical Wavelet Estimation?

Attenuation compensation in viscoacoustic reserve-time migration Jianyong Bai*, Guoquan Chen, David Yingst, and Jacques Leveille, ION Geophysical

From PZ summation to wavefield separation, mirror imaging and up-down deconvolution: the evolution of ocean-bottom seismic data processing

Source function estimation in extended full waveform inversion

Matrix formulation of adjoint Kirchhoff datuming

Th P7 02 A Method to Suppress Salt-related Converted Wave Using 3D Acoustic Modelling

Subsalt imaging by common-azimuth migration

Elastic least-squares reverse time migration using the energy norm Daniel Rocha & Paul Sava Center for Wave Phenomena, Colorado School of Mines

Drift time estimation by dynamic time warping

Complex-beam Migration and Land Depth Tianfei Zhu CGGVeritas, Calgary, Alberta, Canada

Marchenko redatuming: advantages and limitations in complex media Summary Incomplete time reversal and one-sided focusing Introduction

Anisotropic tomography for TTI and VTI media Yang He* and Jun Cai, TGS

Stanford Exploration Project, Report 105, September 5, 2000, pages

2011 SEG SEG San Antonio 2011 Annual Meeting 771. Summary. Method

TOM 2.6. SEG/Houston 2005 Annual Meeting 2581

Solving physics pde s using Gabor multipliers

Wave-equation tomography for anisotropic parameters

Unravel Faults on Seismic Migration Images Using Structure-Oriented, Fault-Preserving and Nonlinear Anisotropic Diffusion Filtering

An example of inverse interpolation accelerated by preconditioning

Residual Statics using CSP gathers

Statics preserving projection filtering Yann Traonmilin*and Necati Gulunay, CGGVeritas

Geophysical Journal International

Full Waveform Inversion (FWI) with wave-equation migration. Gary Margrave Rob Ferguson Chad Hogan Banff, 3 Dec. 2010

Minimum entropy deconvolution with frequency-domain constraints

The Estimation of PS Wave Statics By Using SCDCC + Stack Power Method

Downloaded 08/20/14 to Redistribution subject to SEG license or copyright; see Terms of Use at

Downloaded 05/01/17 to Redistribution subject to SEG license or copyright; see Terms of Use at

Surface multiple attenuation in shallow water, case study on data from the Bonaparte Basin, Australia

Transcription:

GEOPHSICS, VOL. 74, NO. 1 JANUAR-FEBRUAR 29 ; P. V25 V33, 17 FIGS. 1.119/1.343447 Adaptive multiple subtraction using regularized nonstationary regression Sergey Fomel 1 ABSTRACT Stationary regression is the backbone of seismic data-processing algorithms including match filtering, which is commonly applied for adaptive multiple subtraction. However, the assumption of stationarity is not always adequate for describing seismic signals. I have developed a general method of nonstationary regression and that applies to nonstationary match filtering. The key idea is the use of shaping regularization to constrain the variability of nonstationary regression coefficients. Simple computational experiments demonstrate advantages of shaping regularization over classic Tikhonov s regularization, including a more intuitive selection of parameters and a faster iterative convergence. Using benchmark synthetic data examples, I have successfully applied this method to the problem of adaptive subtraction of multiple reflections. INTRODUCTION Many natural phenomena, including geologic events and geophysical data, are fundamentally nonstationary. They might exhibit stationarity on a short timescale but eventually alter their behavior in space and time. Nonstationary adaptive filtering is a well-developed field in signal processing Haykin, 21. In seismic signal processing, nonstationary filters were analyzed by Margrave 1998 and applied to many important problems including multiple suppression Rickett et al., 21, data interpolation Crawley et al., 1999; Curry, 23, and migration deconvolution Guitton, 24; Valenciano et al., 26. In this paper, I present a general approach to designing nonstationary operators, including the case of nonstationary matching and prediction-error filters. The key idea is the application of shaping regularization Fomel, 27b for constraining the continuity and smoothness of the filter coefficients. Regularization makes the estimation problem well-posed and leads to fast numerical algorithms. Advantages of shaping regularization in comparison with the classic Tikhonov regularization include an easier control of the regularization parameters and a faster iterative convergence resulting from better conditioning of the inverted matrix. Adaptive subtraction is a method for matching and removing coherent noise, such as multiple reflections, after their prediction by a data-driven technique Verschuur et al., 1992. Adaptive subtraction involves a matching filter to compensate for the amplitude, phase, and frequency distortions in the predicted noise model. Techniques for matching filtering and adaptive subtraction have been developed and discussed by a number of authors Verschuur et al., 1992; Monk, 1993; Spitz, 1999; van Borselen et al., 23; Wang, 23; Guitton and Verschuur, 24; Abma et al., 25; Lu and Mao, 25; Denisov et al., 26. The regularized nonstationary regression technique proposed in this paper allows the matching filter to become smoothly nonstationary without the need to break the input data into local windows. The paper is organized as follows: I begin with an overview of stationary and nonstationary regression theory and introduce a method of regularized nonstationary regression. Next, I demonstrate this method using toy examples of line fitting and nonstationary deconvolution. Finally, I apply it to the adaptive multiple suppression problem and test its performance using benchmark synthetic data examples. STATIONAR AND NONSTATIONAR REGRESSION Consider a master signal m x, where x represents the coordinates of a multidimensional space, and a collection of slave signals s k x, k 1,2,...,N. The goal of stationary regression is to estimate coefficients a k, k 1,2,...,N so that the prediction error e x m x a k s k x N k 1 1 is minimized in the least-squares sense. Particular examples include: Manuscript received by the Editor 27 July 28; published online 29 December 28. 1 The University of Texas at Austin, Bureau of Economic Geology, John A. and Katherine G. Jackson School of Geosciences, Austin, Texas, U.S.A. E-mail: sergey.fomel@beg.utexas.edu. 29 Society of Exploration Geophysicists. All rights reserved. V25

V26 Fomel Line fitting. Let x be one-dimensional denoted by x, N 2, s 1 x 1, and s 2 x x. The problem of minimizing e x amounts to fitting a straight line a 1 a 2 x to the master signal. Match filtering. If the slave signals are translates of the same signal s x, the regression problem corresponds to the problem of match filtering between s x and m x. In the 1D case, one can take, for example, s k x s x k N/2, which turns the sum in equation 1 into a convolution with an unknown matching filter. Prediction-error filtering. If the slave signals are causal translates of the master signal, the regression problem corresponds to the problem of autoregressive prediction-error filtering. In the 1D case, one can take, for example, s k x m x k Nonstationary regression Nonstationary regression uses a definition similar to equation 1 but allows the coefficients a k to change with x. The error turns into e x m x b k x s k x, N k 1 and the problem of its minimization becomes ill-posed because one can obtain more unknown variables than constraints. The remedy is 9. 8. 7. 6. 5. 4. 3. 2. 9. 8. 7. 6. 5. 4. 3. 2. () = a+b* 5 1 15 2 25 3 35 4 45 5 () = a+b* 5 1 15 2 25 3 35 4 45 5 Figure 1. Line fitting with stationary regression works well for stationary data a but poorly for nonstationary data b. 2 to include additional constraints that limit the allowed variability of the estimated coefficients. The classic regularization method is Tikhonov s regularization Tikhonov, 1963; Engl et al., 1996, which amounts to minimization of the following functional: F b e x 2 2 D b k x 2, N k 1 where D is the regularization operator such as the gradient or Laplacian filter and is a scalar regularization parameter. If D is a linear operator, least-squares estimation reduces to linear inversion b A 1 d, where b b 1 x b 2 x b N x T, d s 1 x m x s 2 x m x s N x m x T, and the elements of matrix A are A ij x s i x s j x 2 ij D T D. Shaping regularization Fomel, 27b formulates the problem differently. Instead of specifying a penalty roughening operator D, one specifies a shaping smoothing operator S. The regularized inversion takes the form b  1 dˆ, where dˆ S s 1 x m x S s 2 x m x S s N x m x T, the elements of matrix  are  ij x 2 ij S s i x s j x 2 ij, and is a scaling coefficient. The main advantage of this approach is the relative ease of controlling the selection of and S in comparison with and D. In all examples in this paper, I define S as Gaussian smoothing with an adjustable radius and choose to be the median value of s i x. As demonstrated in the next section, the matrix  is typically much better conditioned than matrix A, which leads to fast inversion with iterative algorithms. In the case of N 1 regularized division of two signals, a similar construction was applied before to define local seismic attributes Fomel, 27a. TO EAMPLES In this section, I illustrate the general method of regularized nonstationary regression using simple examples of nonstationary line fitting and nonstationary deconvolution. Nonstationary line fitting Figure 1a shows a classic example of linear regression applied as a line-fitting problem. When the same technique is applied to data with nonstationary behavior Figure 1b, stationary regression fails to 3 4 5 6 7

Adaptive multiple suppression V27 produce an accurate fit and creates regions of consistent overprediction and underprediction. One remedy is to extend the model by including nonlinear terms Figure 2a ; another is to break the data into local windows Figure 2b. Both solutions work to a certain extent but are not completely satisfactory because they decrease the estimation stability and introduce additional nonintuitive parameters. The regularized nonstationary solution, defined in the previous section, is shown in Figure 3. When using shaping regularization with smoothing as the shaping operator, the only additional parameter is the radius of the smoothing operator. This toy example makes it easy to compare shaping regularization with the more traditional Tikhonov regularization. Figures 4 and 5 show inverted matrix A from equation 5 and the distribution of its eigenvalues for two values of Tikhonov s regularization parameter, which correspond to mild and strong smoothing constraints. The operator D in this case is the first-order difference. Correspondingly, Figures 6 and 7 show matrix  from equation 7 and the distribution of its eigenvalues for mild and moderate smoothing implemented with shaping. The operator S is Gaussian smoothing controlled by the smoothing radius. When a matrix operator is inverted by an iterative method such as conjugate gradients, two characteristics control the number of itera- 9. 8. 7. 6. 5. 4. 3. 2. () = a()+b()* 5 1 15 2 25 3 35 4 45 5 Figure 3. Nonstationary line fitting by regularized nonstationary regression. 1 2 Column 2 4 6 8 1 3 9. 8. 7. 6. 5. 4. 3. 2. 9. 8. 7. 6. 5. 4. 3. 2. 2 () = a+b*+c* 5 1 15 2 25 3 35 4 45 5 ()=a+b* k k k 5 1 15 2 25 3 35 4 45 5 Figure 2. Nonstationary line fitting using nonlinear terms a and local windows b. Row Magnitude 4 5 6 7 8 9 1 4 3 2 1.5.5 1 2 Eigenvalue number 2 4 6 8 1 Figure 4. Matrix inverted in Tikhonov s regularization applied to nonstationary line fitting a and the distribution of its eigenvalues b. The regularization parameter.1 corresponds to mild smoothing. The condition number is 888888.

V28 Fomel tions and therefore the cost of inversion Golub and Van Loan, 1996; van der Vorst, 23 : 1 2 The condition number the ratio between the largest and the smallest eigenvalue the clustering of eigenvalues Large condition numbers and poor clustering lead to slow convergence. Figures 4 7 demonstrate that both the condition number and the clustering of eigenvalues are significantly better in the case of shaping regularization than in the case of Tikhonov s regularization. In toy problems, this difference in behavior is not critical, because one can easily invert both matrices exactly. However, this difference becomes important in large-scale applications when inversion is iterative and saving the number of iterations is crucial for performance. As the smoothing radius increases, matrix  approaches the identity matrix and the result of nonstationary regression regularized by shaping approaches the result of stationary regression. This intuitively pleasing behavior is difficult to emulate with Tikhonov s regularization. Nonstationary deconvolution Figure 8 shows an application of regularized nonstationary regression to a benchmark deconvolution test from Claerbout 28. The input signal is a synthetic trace that contains events with variable frequencies. A prediction-error filter is estimated by setting N 2, s 1 x m x 1, and s 2 x m x 2. I use Gaussian smoothing with a five-sample radius as the shaping operator S. The deconvolved signal bottom plot in Figure 8 shows the nonstationary reverberations correctly deconvolved. Column 2 4 6 8 1 Column 2 4 6 8 1 1 1 2 2 3 3 4 4 Row 5 Row 5 6 6 7 7 8 8 9 9 1 1-1 1 2.5 1 4 Eigenvalue number 2 4 6 8 1 Eigenvalue number 2 4 6 8 1 35 3. 3 25 Magnitude 2 15 Magnitude 2. 1 5.5. Figure 5. Matrix inverted in Tikhonov s regularization applied to nonstationary line fitting a and the distribution of its eigenvalues b. The regularization parameter 1 corresponds to strong smoothing. The condition number is 1473. Eigenvalues are poorly clustered. Figure 6. Matrix inverted in shaping regularization applied to nonstationary line fitting a and the distribution of its eigenvalues b. The smoothing radius is three samples mild smoothing. The condition number is 655.

Adaptive multiple suppression V29 A three-point prediction-error filter 1,a 1,a 2 can predict an attenuating sinusoidal signal m x x cos x, where is frequency and is the exponent factor, provided the filter coefficients are defined as a 1 2 cos ; a 2 2. This fact follows from the simple identity m x a 1 m x 1 a 2 m x 2 x cos x 8 9 1 2 cos cos x cos x 2. 11 Column 2 4 6 8 1 According to equations 9 and 1, one can obtain an estimate of the local frequency x from the nonstationary coefficients b 1 x a 1 x and b 2 x a 2 x as follows: x arccos b 1 x 2 b2 x. 12 Figure 9 shows the frequency of the components in the input nonstationary signal from Figure 8 and the local frequency estimate using equation 12. A reasonably accurate match between the true nonstationary frequencies and their estimate from nonstationary regression can be observed. The local frequency attribute Fomel, 27a provides a different approach to the same problem. ADAPTIVE MULTIPLE SUBTRACTION In this section, I apply regularized nonstationary regression to the problem of adaptive subtraction and evaluate its performance using benchmark synthetic tests. 1 2 3 Input 4 Row 5 6 7 8 9 1 Deconvolution Magnitude 2.8 2.6 2.4 2.2 2. 1.8 1.6 1.4 1.2.8.6.4.2..2.4.6.8 1 Eigenvalue number 2 4 6 8 1 Figure 7. Matrix inverted in shaping regularization applied to nonstationary line fitting a and the distribution of its eigenvalues b. The smoothing radius is 15 samples moderate smoothing. The condition number is 26. Eigenvalues are well clustered. 5 1 15 2 25 3 35 4 Sample Figure 8. Benchmark test of nonstationary deconvolution from Claerbout 28. Top: input signal, bottom: deconvolved signal using nonstationary regularized regression. Frequency (cycles).9.8.7.6.5.4.3.2.1. 5 1 15 2 25 3 35 4 Sample Figure 9. Frequency of components in the input synthetic signal from Figure 8 solid line and the local frequency estimate from nonstationary deconvolution dashed line.

V3 Fomel One possible variation of nonstationary regression applied to adaptive subtraction is an introduction of the signal-prediction filter Spitz, 1999. The signal prediction applies to the residual error e x in equation 2, thus modifying the objective function Luo et al., 23; Guitton, 25. Abma test Abma et al. 25 present a comparison of adaptive subtraction algorithms used for multiple suppression. I used Abma s benchmark examples to illustrate an application of nonstationary regression to the adaptive subtraction problem. Figure 1 shows the first test: the input data Figure 1a contain a horizontal signal event and dipping noise events. We are provided with a model of the noise Figure 1b ; however, the model events are slightly shifted and have a different amplitude behavior. This test simulates a typical situation in adaptive surface-related multiple elimination, in which there are phase and amplitude distortions in the multiple model caused by an imperfect prediction. To handle the variability in noise amplitudes, I designed a nonstationary matching filter with 13 coefficients and three-sample smoothing radius and applied it to match the noise model to the data. Subtracting matched noise Figure 1d produces the desired and nearly perfect signal estimate Figure 1c. The variability of filter coefficients is illustrated in Figure 11, which displays the zero-lag coefficient and the mean coefficient of the nonstationary matching filter. Another benchmark from Abma et al. 25 is shown in Figure 12. This time, the noise part of the data is a curved event that has a Figure 1. Benchmark test on adaptive subtraction from Abma et al. 25. a Input synthetic data. b Model of the noise containing amplitude and phase differences with respect to the corresponding part of the data. c Extracted signal. d Estimated noise.. 2 4 6 8 1 12. 2 4 6 8 1 12.1.1.2.2.3.3.4.4.5.5.6.6.7.7.8.8.9.9 Data Multiple model. 2 4 6 8 1 12. 2 4 6 8 1 12.1.1.2.2.3.3.4.4.5.5.6.6.7.7.8.8.9.9 Estimated signal Matched multiples Figure 11. Variability of nonstationary match filter coefficients for the example shown in Figure 1. a Zero-lag coefficient. b Mean coefficient.

Adaptive multiple suppression V31 predictive model Figure 12b with some amplitude and phase differences. Nonstationary regularized regression correctly predicts the noise signal Figure 12d using match filtering and produces an accurate signal estimate Figure 12c. The variability of nonstationary filter coefficients is shown in Figure 13. Spitz test The next benchmark example Figure 14 reproduced a test created by Spitz 27, which simulates a seismic gather after migration or normal moveout correction. The signal primary reflection event is horizontal, whereas the noise multiple reflection event exhibits curvature and overlaps with the signal at small offsets Figure 14a. Multiple prediction Figure 14b contains a curved event but with incorrect curvature. As in the previous examples, nonstationary regularized regression correctly predicted the noise signal Figure 14d using match filtering and produces an accurate signal estimate Figure 14c. Pluto test Finally, Figure 15 shows an application of the nonstationary matching technique to the Pluto synthetic data set, a well-known benchmark for adaptive multiple subtraction. Matching and subtracting an imperfect model of the multiples created by the surfacerelated multiple elimination approach of Verschuur et al. 1992 leaves a clean estimate of primary reflections. Figure 16 shows a comparison between the multiple model obtained by surface-related prediction and the multiple model generated by nonstationary. 2 4 6 8 1 12.1.2.3.4.5.6.7.8.9 Data. 2 4 6 8 1 12.1.2.3.4.5.6.7.8.9 Estimated signal. 2 4 6 8 1 12.1.2.3.4.5.6.7.8.9 Multiple model. 2 4 6 8 1 12.1.2.3.4.5.6.7.8.9 Matched multiples Figure 12. Benchmark test on adaptive subtraction from Abma et al. 25. a Input synthetic data. b Model of the noise containing amplitude and phase differences with respect to the corresponding part of the data. c Extracted signal. d Estimated noise. Figure 13. Variability of nonstationary match filter coefficients for the example shown in Figure 12. a Zero-lag coefficient. b Mean coefficient.

V32 Fomel Figure 14. Benchmark test on multiple attenuation from Spitz 27. a Input synthetic data containing signal primary reflection and noise multiple reflection events. b Model of the noise containing phase differences with respect to the corresponding part of the data. c Extracted signal. d Estimated noise..45.5.55.6.65.7.75.8.45.5.55.6.65.7.75 Offset (km) Offset (km).1.2.3.4.1.2.3.4.45.5.55.6.65.7.75.8 Data Multiple model Offset (km) Offset (km).1.2.3.4.1.2.3.4.45.5.55.6.65.7.75.8 Estimated primary.8 Estimated multiple..5 2. 3. 4. 5. 6. 1 2 3 4 5 6 7 8. 1 2 3 4 5 6 7 8.5 2. 3. 4. 5. 6. Figure 15. Adaptive multiple subtraction in the Pluto synthetic data set. a Input data. b Extracted signal. Surface-related multiples are successfully subtracted. 1 2 3 4 5 6 7 8..5 2. 3. 4. 5. 6.. 1 2 3 4 5 6 7 8.5 2. 3. 4. 5. 6. Figure 16. Multiple model from surface-related prediction a and estimated multiples b for the Pluto synthetic data set.

Adaptive multiple suppression V33 Figure 17. Variability of nonstationary match filter coefficients for the Pluto test. a Zero-lag coefficient. b Mean coefficient. matching. The matching filter nonstationarity is depicted in Figure 17, which shows the variability of filter coefficients with time and space. CONCLUSIONS I have presented a general method for regularized nonstationary regression. The key idea is the application of shaping regularization for constraining the variability of nonstationary coefficients. Shaping regularization has clear advantages over the conventional Tikhonov regularization: a more intuitive selection of regularization parameters and a faster iterative convergence because of better conditioning and eigenvalue clustering of the inverted matrix. I have shown examples of applying regularized regression to benchmark tests of adaptive multiple subtraction. When the signal characteristics change with time or space, regularized nonstationary regression provides an effective description of the signal. This approach does not require breaking the input data into local windows, although it is analogous conceptually to sliding maximum overlap spatial-temporal windows. ACKNOWLEDGMENTS I thank StatoilHydro for a partial financial support of this work. I would like to thank Dave Hale, Long Jin, Dmitry Lokshtanov, and Gary Sitton for inspiring discussions and Ray Abma for providing benchmark data sets for adaptive subtraction. The Pluto synthetic was generated by the SMAART joint venture. This publication is authorized by the Director, Bureau of Economic Geology, The University of Texas at Austin. REFERENCES Abma, R., N. Kabir, K. H. Matson, S. Michell, S. Shaw, and B. McLain, 25, Comparisons of adaptive subtraction methods for multiple attenuation: The Leading Edge, 24, 277 28. Claerbout, J., 28, Image estimation by example: Environmental soundings image enhancement: Stanford Exploration Project. Crawley, S., J. Claerbout, and R. Clapp, 1999, Interpolation with smoothly nonstationary prediction-error filters: 69th Annual International Meeting, SEG, ExpandedAbstracts, 1154 1157. Curry, W., 23, Interpolation of irregularly sampled data with nonstationary, multiscale prediction-error filters: 73rdAnnual International Meeting, SEG, ExpandedAbstracts, 1913 1916. Denisov, M. S., M. A. Polubojarinov, and D. B. Finikov, 26, Robust methods of adaptive seismic multiple suppression: 68th Conference and Exhibition, EAGE, ExpandedAbstracts, B33. Engl, H., M. Hanke, and A. Neubauer, 1996, Regularization of inverse problems: KluwerAcademic Publishers. Fomel, S., 27a, Local seismic attributes: Geophysics, 72, no. 3, A29-A33., 27b, Shaping regularization in geophysical-estimation problems: Geophysics, 72, no. 2, R29-R36. Golub, G. H., and C. F. Van Loan, 1996, Matrix computations: The John Hopkins University Press. Guitton, A., 24, Amplitude and kinematic corrections of migrated images for nonunitary imaging operators: Geophysics, 69, 117 124., 25, Multiple attenuation in complex geology with a pattern-based approach: Geophysics, 7, no. 4, V97-V17. Guitton, A., and D. Verschuur, 24, Adaptive subtraction of multiples using the L1-norm: Geophysical Prospecting, 52, 27 38. Haykin, S., 21, Adaptive filter theory: Prentice Hall. Lu, W., and F. Mao, 25, Adaptive multiple subtraction using independent component analysis: The Leading Edge, 24, 282 284. Luo,., P. G. Kelamis, and. Wang, 23, Simultaneous inversion of multiples and primaries: Inversion versus subtraction: The Leading Edge, 22, 814 818. Margrave, G. F., 1998, Theory of nonstationary linear filtering in the Fourier domain with application to time-variant filtering: Geophysics, 63, 244 259. Monk, D. J., 1993, Wave-equation multiple suppression using constrained gross-equalization: Geophysical Prospecting, 41, 725 736. Rickett, J., A. Guitton, and D. Gratwick, 21, Adaptive multiple subtraction with non-stationary helical shaping filters: Presented at the 63rd Annual International Meeting, EAGE. Spitz, S., 1999, Pattern recognition, spatial predictability, and subtraction of multiple events: The Leading Edge, 18, 55 58., 27, Perturbed prediction-driven multiple elimination: A PEF scheme: 77th Annual International Meeting, SEG, Expanded Abstracts, 2649 2653. Tikhonov, A. N., 1963, Solution of incorrectly formulated problems and the regularization method: Soviet Mathematics Doklady, 135 138. Valenciano, A. A., B. Biondi, and A. Guitton, 26, Target-oriented waveequation inversion: Geophysics, 71, no. 4, A35-A38. van Borselen, R., G. Fookes, and J. Brittan, 23, Target-oriented adaptive subtraction in data-driven multiple removal: The Leading Edge, 22, 34 343. van dervorst, H. A., 23, Iterative Krylov methods for large linear systems: Cambridge University Press. Verschuur, D. J., A. J. Berkhout, and C. P. A. Wapenaar, 1992, Adaptive surface-related multiple elimination: Geophysics, 57, 1166 1177. Wang,., 23, Multiple subtraction using an expanded multichannel matching filter: Geophysics, 68, 346 354.