Wavelet Extraction: an essay in model selection and uncertainty. James Gunning, CSIRO Petroleum, Australia

Size: px
Start display at page:

Download "Wavelet Extraction: an essay in model selection and uncertainty. James Gunning, CSIRO Petroleum, Australia"

Transcription

1 Wavelet Extraction: an essay in model selection and uncertainty James Gunning, CSIRO Petroleum, Australia

2 Outline Model selection, inference and prediction: Generalities Wavelet extraction Model formulation Generic inversion/inference features for single models Model choice scenarios. 3 examples Morals & conclusions

3 Model selection Classic statistics problem line? or parabola? or sextic? y Foundations of science Newton or Ptolemy? x F = m a

4 Generalities Geostatistics is only one branch of multivariate statistics. Every result of significance can be derived from conventional multivariate estimation theory. A Bayesian spin is helpful. The rest of the world understands us better if we write things this way. As statisticians, we have the duties of Model inference Significance testing Model refinement/testing Making predictions with error estimates

5 Model selection & Bayes factors Suppose there are k=1 N models to explain data D, with parameters m k The model has prior P(m k k)p(k), likelihood L(D m k,k) Here s Three very useful things: Marginal model likelihood: P(D k)= L(D m k,k)p(m k k) dm k Bayes factor B ij = P(D i)/ P(D j) Model probability of model k: P(k D) = N P(D k)p(k) j=1 P(D j)p( j)

6 Predictive distributions Bayesian model averaging: prediction for new y P(y D) = k P(y M k,d) P(D k) P(y D) Mixture distributions Models/rock-types/properties Most helpful way to manage heavy tails Often: uncertainties within a model < uncertainty between models y

7 Noise ε (D-f(m)) ε= ε meas +ε model Noise estimation Often ε model > ε meas : we re prepared to live with approximate models ε meas usually stationary. Maybe biased. Roughly known. ε model often nonstationary. Correlated. Often biased. Usually unknown. We blaze away: {ε, σ 2 } ~ N(0, σ 2 )P(σ 2 ) Inference of p(ε) chief issue p(ε) strongly coupled to model dimensionality Demands sensible priors in Bayesian formulation Simplest form adds pieces like - Log(Π) ~ D-f(m) 2 /(2σ 2) +n D log(σ 2 )/2 - Absolute statements of model significance no longer possible

8 Marginal model likelihood estimators Full linearisation+conjugate priors, analytical Quadratures. Too hard as d=dim(m) gets large MCMC Harmonic mean averages: P(D k) = L(D m k,k)p(m k,k)dm k (1) ˆ P (D k) = Unstable as Nsamples Laplace-Metropolis approx. P(D k) 1 L(D m k,k) Π(mk D )(MCMC ) BIC asymptotics (eqn (1) does the overfit penalty automatically) 1 e f (m) dm (2π) d /2 det{ 2 f m i m j log(p(d k)) ~ log(l(d ˆ m )P( ˆ m ))) + d 2 ˆ m } 1 e f ( ˆ m ) = (2π) d /2 H 1/2 L(D ˆ log(n D ) + O(n 1/2 ) m )P ( m ˆ )

9 Issues in prior formulation Lindley paradox (1957) Models M 1 and M 2, with d 2 > d 1. Suppose they share (nested) parameters m with prior m 1 ~N(0,σ 2 C 1 ) m 2 ~N({0,0},σ 2 diag{c 1,C 2 }) P(σ)~1/σ (Jeffrey s prior) Then the Bayes factor is (Smith & Spiegelhalter 1980) B 12 ~ C 2 1/2 X 2 T X 2 1/2 (1+((d 2 -d 1 )/(n-d 1 )F) n/2 Always favours simplest model (1) as C 2 (even for proper prior on m 2!) Setting of reasonable prior variances very important in comparing different dimensional models

10 Segmentation/Blocking algorithms Changepoints a classic modelchoice problem Maximum likelihood configurations for k changepoints τ j accessible from dynamic programming O(kn 2 ) (D.M.Hawkins, Comp. Stat. & Data Analysis 37(3), 2001) Uses: Thinning reflectivity sequence Optimal checkshot choices log(l(y,m,τ)) = 1 2 k τ j j=1 i=τ j 1 +1 (y i m j ) 2 /σ 2 + n 2 log(σ 2 ) + BIC

11 Wavelet extraction problem True-amplitude imaged seismic t (time) wavelet Systematic registration error? CONVOLUTION Z (depth) reflections noise level from well logs

12 Real wavelets and effective wavelets: Ricker 1953 (dynamite)

13 Motivations Wavelet extraction a critical process in inversion studies Commercial codes don't (or wont) tell you Noise estimates Non-normal-incidence wavelets Time to depth or positioning errors May struggle with Multiple wells Deviated wells wavelet uncertainties

14 Wishlist Wavelet coefficient extraction (plus errors) Wavelet length estimation (plus errors) Noise estimation (plus errors) Petrophysics model choices Time to depth adjustments(plus errors) Integrates checkshots and sonic logs Positioning adjustments (plus errors) Multi-well Multi-stack

15 Wavelet Parameterization (1) Wavelet coefficients m w Nyquist-rate knots on cubic splines with zero endpoints & derivatives Ensures decent tapering & correct bandwidth Span? (see later) Prior P(m w )=N(0,s 2 I)N(t p,σ p2 )N(Φ(ω),σ φ2 I) Optional timing prior Optional phase prior in passband

16 Checkshot Parameterization (2) Time to depth parameters t N(t m,i,σ m,i2 ) N(t c,i,σ c,i2 ) Markers (geological identifications) Checkshots (e.g. from VSP tools) Global (well specific) time shift t g ±σ tg z v int (z i+1 +1-z i,t i+1 -t i ) ~ N( N(v log-average average, σ v2 )

17 Petrophysics parameterizations: Multi stack Near stack Far stack

18 Petrophysics Parameterizations(4): AVO effects (Linearized Zoeppritz) General approximate p-p reflection coefficient normal incidence angle unknown? R=½( ρ/ ρ + v p / v p ) + B( θ 2 (½ v p / v p -2v s2 ( ρ/ρ+2 v s /v s )/v p2 ) + θ 2 ) δ/2 shear dependent Thomsen anisotropy parameters: Stack angle θ 2 = v p2 /(V stack4 T stack2 /<X stack2 >) R computed at block boundaries: v p,v s,ρ are Backusupscaled segment properties (i.e. error-free log data) Anisotropy: δ Ν(δ f,σ f2 ), f=lumped facies parameter label (from classification of segment)

19 Anisotropy issues Processing issues for AVO Many potential amplitude effects at wider offset Absorption/scattering/spreading/acquisition footprints, etc.. etc.. etc Ch. 4.3 Quantitative seismic interpretation Avseth, Mukerji, Mavko.

20 Preprocessing: Impedance Blocking Based on segmentation of p- wave impedance (ρv p ) Max Lik. methods O(N 2 ), or Blended hierarchical stepwise segment/aggregate method (O(Nlog(k)) ref: D.M.Hawkins, Comp. Stat. & Data Analysis 37(3), 2001 Increases speed

21 Optimisation of Bayesian Posterior m={m wavelet,m checkshot,m registration,m petrophysics } Maximise exp(-χ 2 /2 ) χ 2 = Σ t,stacks,wells (w(m)*r eff S) 2 /σ s 2 "good synthetic" + (d+1) Σ log(σ stacks s ) "noise normalisation + Σ intervals,wells (v int (m)-<v int >)2 /σ v 2 prior: interval velocities c.f. sonic logs + (m-<m>) T C -1 p (m-<m>) "prior on checkshots, wavelet coeffs + wavelet timing/phase prior "e.g. constant phase"

22 Optimisation, Hessians, Approximate uncertainties parameter 2 3 Maximum aposteriori estimates Quadratic approximation to uncertainties -(1) direct from F.D. Hessian or -(2) from Bayes' update formula Gauss Newton optimiser log(posterior) surface - C=(Cp -1 +X T C D -1 X) -1 parameter 1 Gauss-Newton, BFGS methods: see e.g. Nocedal & Wright "Numerical Optimization" O(d 2 ) x O(forward model cost) x N(models) Typical dimensionalities: d=10-100, n =

23 Optimisation outputs Maximum aposteriori parameters SU wavelets, ascii time-to-depth files, visualization dumps etc. Covariance diagonals (parameter uncertainties) Posterior model probabilities (wavelet span, checkshot choice etc) stochastic wavelets from the posterior

24 Generic aspects of well-ties Statistically insignificant realisations Statistically insignificant Typical commercial extraction Bayes ML wavelet

25 Joint straight-hole & sidetrack No logs: checkshot uncertainties go to prior Looser checkshots with weaker amplitudes Tight checkshot uncertainties at strong amplitudes tvd Sidetrack tvd Straight

26 Straight hole detail with logs Interval sonic velocities constrained to within 5% of log average

27 Parameter uncertainty cross sections log(posterior) log(posterior) wavelet coefficients One standard deviation checkshot knots repulsion to keep interval velocities sane

28 General effect of allowing more freedom Wavelets sharpen, noise reduces

29 Cross well issues: Wavelets and Ties Synthetics/traces from joint extraction Wavelets from various extractions

30 Relations to Cross-validation Leave-one-out pseudo Bayes factor B 12(cross val ) = n D p(d i d { j i},m 1 ) p(d i d { j i},m 2 ) i=1 Has asymptotics of Akaike information criterion Log(P(D k)) += d (c.f. Bayes info. crit. = (d/2)log(n D ) (less severe on richer models)

31 3 Model Selection examples

32 (1) Simple Wavelet-length choice

33 Absolute significance tests MAP Wavelet length not shortest model MAP noise level exceptional on null-hypothesis test Perform ensemble of extractions on fake seismic data with matched bandwidth and univariate statistics: P(σ null ) σ true σ null

34 (2) Checkshot choice (25 models) Changepoint cascade

35 Richest model

36 Max. aposteriori model

37 Model probabilities and noise

38 (3) Naïve facies-driven anisotropy choice δ~0.15

39 Conclusions Model selection is desirable in nearly all inverse problems Marginal model likelihoods can be tricky Careful construction of prior Efficient (differentiable) forward models Efficient optimisers Bayesian model selection is quite opinionated. Very Occamist: severe on overparameterised models Selection is a doable problem for wavelet extraction. Produces Often compact wavelets Simplified checkshots Significant potential for missing-data (petrophysics) problems

40 O me, the word 'choose!' I may neither choose whom I would nor refuse whom I dislike. Portia, The Merchant of Venice. More info: Wavelet extractor open source code (java), demos, papers. Part of Delivery suite (Bayesian seismic inversion) Computers and Geosciences 32 (2006)

41 Data-positioning uncertainty (1) y(m) = f(m) an unusual regression problem Simple toy problem; fitting a line y=a+bx: y ~ X.m + e, Jeffrey's prior on e ~ N(0,σ) Priors: P(σ ) ~ 1/ σ P(m σ) ~ N(0,g σ 2 (X T X) -1 ) ("Zellner") Posterior marginal = L(y m,σ)p(m σ)p(σ)dmdσ ~ (1+g(1-R 2 )) -(n-1)/2 Where R 2 is usual coeff. of determination For two samples of random y B 12 ~ ((1+g(1-R 12 ))/ (1+g(1-R 22 ))) -(n-1)/2

42 Data-positioning uncertainty (2) Fluctuation in B 12 for problem with y=1+x+(25% fixed noise) + (5% fluctuations) frequency log 10 (BF 12 )

Delivery: an open-source Bayesian seismic inversion tool. James Gunning, CSIRO Petroleum Michael Glinsky,, BHP Billiton

Delivery: an open-source Bayesian seismic inversion tool. James Gunning, CSIRO Petroleum Michael Glinsky,, BHP Billiton Delivery: an open-source Bayesian seismic inversion tool James Gunning, CSIRO Petroleum Michael Glinsky,, BHP Billiton Application areas Development, appraisal and exploration situations Requisite information

More information

Model Selection for Gaussian Processes

Model Selection for Gaussian Processes Institute for Adaptive and Neural Computation School of Informatics,, UK December 26 Outline GP basics Model selection: covariance functions and parameterizations Criteria for model selection Marginal

More information

Bayesian lithology/fluid inversion comparison of two algorithms

Bayesian lithology/fluid inversion comparison of two algorithms Comput Geosci () 14:357 367 DOI.07/s596-009-9155-9 REVIEW PAPER Bayesian lithology/fluid inversion comparison of two algorithms Marit Ulvmoen Hugo Hammer Received: 2 April 09 / Accepted: 17 August 09 /

More information

Seismic reservoir characterization in offshore Nile Delta.

Seismic reservoir characterization in offshore Nile Delta. Seismic reservoir characterization in offshore Nile Delta. Part II: Probabilistic petrophysical-seismic inversion M. Aleardi 1, F. Ciabarri 2, B. Garcea 2, A. Mazzotti 1 1 Earth Sciences Department, University

More information

7. Estimation and hypothesis testing. Objective. Recommended reading

7. Estimation and hypothesis testing. Objective. Recommended reading 7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing

More information

Model comparison and selection

Model comparison and selection BS2 Statistical Inference, Lectures 9 and 10, Hilary Term 2008 March 2, 2008 Hypothesis testing Consider two alternative models M 1 = {f (x; θ), θ Θ 1 } and M 2 = {f (x; θ), θ Θ 2 } for a sample (X = x)

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 CS students: don t forget to re-register in CS-535D. Even if you just audit this course, please do register.

More information

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 2: Multivariate Gaussians

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 2: Multivariate Gaussians Engineering Part IIB: Module F Statistical Pattern Processing University of Cambridge Engineering Part IIB Module F: Statistical Pattern Processing Handout : Multivariate Gaussians. Generative Model Decision

More information

Stochastic vs Deterministic Pre-stack Inversion Methods. Brian Russell

Stochastic vs Deterministic Pre-stack Inversion Methods. Brian Russell Stochastic vs Deterministic Pre-stack Inversion Methods Brian Russell Introduction Seismic reservoir analysis techniques utilize the fact that seismic amplitudes contain information about the geological

More information

7. Estimation and hypothesis testing. Objective. Recommended reading

7. Estimation and hypothesis testing. Objective. Recommended reading 7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing

More information

STAT Financial Time Series

STAT Financial Time Series STAT 6104 - Financial Time Series Chapter 4 - Estimation in the time Domain Chun Yip Yau (CUHK) STAT 6104:Financial Time Series 1 / 46 Agenda 1 Introduction 2 Moment Estimates 3 Autoregressive Models (AR

More information

Downloaded 10/02/18 to Redistribution subject to SEG license or copyright; see Terms of Use at

Downloaded 10/02/18 to Redistribution subject to SEG license or copyright; see Terms of Use at Multi-scenario, multi-realization seismic inversion for probabilistic seismic reservoir characterization Kester Waters* and Michael Kemper, Ikon Science Ltd. Summary We propose a two tiered inversion strategy

More information

Bayesian inference J. Daunizeau

Bayesian inference J. Daunizeau Bayesian inference J. Daunizeau Brain and Spine Institute, Paris, France Wellcome Trust Centre for Neuroimaging, London, UK Overview of the talk 1 Probabilistic modelling and representation of uncertainty

More information

Lecture 13: Data Modelling and Distributions. Intelligent Data Analysis and Probabilistic Inference Lecture 13 Slide No 1

Lecture 13: Data Modelling and Distributions. Intelligent Data Analysis and Probabilistic Inference Lecture 13 Slide No 1 Lecture 13: Data Modelling and Distributions Intelligent Data Analysis and Probabilistic Inference Lecture 13 Slide No 1 Why data distributions? It is a well established fact that many naturally occurring

More information

g-priors for Linear Regression

g-priors for Linear Regression Stat60: Bayesian Modeling and Inference Lecture Date: March 15, 010 g-priors for Linear Regression Lecturer: Michael I. Jordan Scribe: Andrew H. Chan 1 Linear regression and g-priors In the last lecture,

More information

Consistent Downscaling of Seismic Inversions to Cornerpoint Flow Models SPE

Consistent Downscaling of Seismic Inversions to Cornerpoint Flow Models SPE Consistent Downscaling of Seismic Inversions to Cornerpoint Flow Models SPE 103268 Subhash Kalla LSU Christopher D. White LSU James S. Gunning CSIRO Michael E. Glinsky BHP-Billiton Contents Method overview

More information

Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv: )

Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv: ) Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv:1809.05778) Workshop on Advanced Statistics for Physics Discovery aspd.stat.unipd.it Department of Statistical Sciences, University of

More information

SRNDNA Model Fitting in RL Workshop

SRNDNA Model Fitting in RL Workshop SRNDNA Model Fitting in RL Workshop yael@princeton.edu Topics: 1. trial-by-trial model fitting (morning; Yael) 2. model comparison (morning; Yael) 3. advanced topics (hierarchical fitting etc.) (afternoon;

More information

Machine Learning - MT Classification: Generative Models

Machine Learning - MT Classification: Generative Models Machine Learning - MT 2016 7. Classification: Generative Models Varun Kanade University of Oxford October 31, 2016 Announcements Practical 1 Submission Try to get signed off during session itself Otherwise,

More information

Physics 403. Segev BenZvi. Numerical Methods, Maximum Likelihood, and Least Squares. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Numerical Methods, Maximum Likelihood, and Least Squares. Department of Physics and Astronomy University of Rochester Physics 403 Numerical Methods, Maximum Likelihood, and Least Squares Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Quadratic Approximation

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

Stat260: Bayesian Modeling and Inference Lecture Date: February 10th, Jeffreys priors. exp 1 ) p 2

Stat260: Bayesian Modeling and Inference Lecture Date: February 10th, Jeffreys priors. exp 1 ) p 2 Stat260: Bayesian Modeling and Inference Lecture Date: February 10th, 2010 Jeffreys priors Lecturer: Michael I. Jordan Scribe: Timothy Hunter 1 Priors for the multivariate Gaussian Consider a multivariate

More information

Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis

Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis Jeffrey S. Morris University of Texas, MD Anderson Cancer Center Joint wor with Marina Vannucci, Philip J. Brown,

More information

CSci 8980: Advanced Topics in Graphical Models Gaussian Processes

CSci 8980: Advanced Topics in Graphical Models Gaussian Processes CSci 8980: Advanced Topics in Graphical Models Gaussian Processes Instructor: Arindam Banerjee November 15, 2007 Gaussian Processes Outline Gaussian Processes Outline Parametric Bayesian Regression Gaussian

More information

Bayesian Lithology-Fluid Prediction and Simulation based. on a Markov Chain Prior Model

Bayesian Lithology-Fluid Prediction and Simulation based. on a Markov Chain Prior Model Bayesian Lithology-Fluid Prediction and Simulation based on a Markov Chain Prior Model Anne Louise Larsen Formerly Norwegian University of Science and Technology, N-7491 Trondheim, Norway; presently Schlumberger

More information

Penalized Loss functions for Bayesian Model Choice

Penalized Loss functions for Bayesian Model Choice Penalized Loss functions for Bayesian Model Choice Martyn International Agency for Research on Cancer Lyon, France 13 November 2009 The pure approach For a Bayesian purist, all uncertainty is represented

More information

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall Machine Learning Gaussian Mixture Models Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall 2012 1 The Generative Model POV We think of the data as being generated from some process. We assume

More information

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis Lecture 3 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,

More information

Gaussian processes for spatial modelling in environmental health: parameterizing for flexibility vs. computational efficiency

Gaussian processes for spatial modelling in environmental health: parameterizing for flexibility vs. computational efficiency Gaussian processes for spatial modelling in environmental health: parameterizing for flexibility vs. computational efficiency Chris Paciorek March 11, 2005 Department of Biostatistics Harvard School of

More information

When one of things that you don t know is the number of things that you don t know

When one of things that you don t know is the number of things that you don t know When one of things that you don t know is the number of things that you don t know Malcolm Sambridge Research School of Earth Sciences, Australian National University Canberra, Australia. Sambridge. M.,

More information

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception LINEAR MODELS FOR CLASSIFICATION Classification: Problem Statement 2 In regression, we are modeling the relationship between a continuous input variable x and a continuous target variable t. In classification,

More information

The Laplace Approximation

The Laplace Approximation The Laplace Approximation Sargur N. University at Buffalo, State University of New York USA Topics in Linear Models for Classification Overview 1. Discriminant Functions 2. Probabilistic Generative Models

More information

GWAS IV: Bayesian linear (variance component) models

GWAS IV: Bayesian linear (variance component) models GWAS IV: Bayesian linear (variance component) models Dr. Oliver Stegle Christoh Lippert Prof. Dr. Karsten Borgwardt Max-Planck-Institutes Tübingen, Germany Tübingen Summer 2011 Oliver Stegle GWAS IV: Bayesian

More information

23855 Rock Physics Constraints on Seismic Inversion

23855 Rock Physics Constraints on Seismic Inversion 23855 Rock Physics Constraints on Seismic Inversion M. Sams* (Ikon Science Ltd) & D. Saussus (Ikon Science) SUMMARY Seismic data are bandlimited, offset limited and noisy. Consequently interpretation of

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population

More information

Geostatistics for Seismic Data Integration in Earth Models

Geostatistics for Seismic Data Integration in Earth Models 2003 Distinguished Instructor Short Course Distinguished Instructor Series, No. 6 sponsored by the Society of Exploration Geophysicists European Association of Geoscientists & Engineers SUB Gottingen 7

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

Gaussian Process Regression

Gaussian Process Regression Gaussian Process Regression 4F1 Pattern Recognition, 21 Carl Edward Rasmussen Department of Engineering, University of Cambridge November 11th - 16th, 21 Rasmussen (Engineering, Cambridge) Gaussian Process

More information

How to build an automatic statistician

How to build an automatic statistician How to build an automatic statistician James Robert Lloyd 1, David Duvenaud 1, Roger Grosse 2, Joshua Tenenbaum 2, Zoubin Ghahramani 1 1: Department of Engineering, University of Cambridge, UK 2: Massachusetts

More information

Thomas Bayes versus the wedge model: An example inference using a geostatistical prior function

Thomas Bayes versus the wedge model: An example inference using a geostatistical prior function Thomas Bayes versus the wedge model: An example inference using a geostatistical prior function Jason M. McCrank, Gary F. Margrave, and Don C. Lawton ABSTRACT The Bayesian inference is used to estimate

More information

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 2: Multivariate Gaussians

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 2: Multivariate Gaussians University of Cambridge Engineering Part IIB Module 4F: Statistical Pattern Processing Handout 2: Multivariate Gaussians.2.5..5 8 6 4 2 2 4 6 8 Mark Gales mjfg@eng.cam.ac.uk Michaelmas 2 2 Engineering

More information

Chris Bishop s PRML Ch. 8: Graphical Models

Chris Bishop s PRML Ch. 8: Graphical Models Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular

More information

GAUSSIAN PROCESS REGRESSION

GAUSSIAN PROCESS REGRESSION GAUSSIAN PROCESS REGRESSION CSE 515T Spring 2015 1. BACKGROUND The kernel trick again... The Kernel Trick Consider again the linear regression model: y(x) = φ(x) w + ε, with prior p(w) = N (w; 0, Σ). The

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

Bayesian methods in economics and finance

Bayesian methods in economics and finance 1/26 Bayesian methods in economics and finance Linear regression: Bayesian model selection and sparsity priors Linear Regression 2/26 Linear regression Model for relationship between (several) independent

More information

Design of Text Mining Experiments. Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt.

Design of Text Mining Experiments. Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt. Design of Text Mining Experiments Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt.taddy/research Active Learning: a flavor of design of experiments Optimal : consider

More information

. Also, in this case, p i = N1 ) T, (2) where. I γ C N(N 2 2 F + N1 2 Q)

. Also, in this case, p i = N1 ) T, (2) where. I γ C N(N 2 2 F + N1 2 Q) Supplementary information S7 Testing for association at imputed SPs puted SPs Score tests A Score Test needs calculations of the observed data score and information matrix only under the null hypothesis,

More information

Empirical comparison of two Bayesian lithology fluid prediction algorithms

Empirical comparison of two Bayesian lithology fluid prediction algorithms Empirical comparison of two Bayesian lithology fluid prediction algorithms Hugo Hammer and Marit Ulvmoen Department of Mathematical Sciences Norwegian University of Science and Technology Trondheim, Norway

More information

PATTERN RECOGNITION AND MACHINE LEARNING

PATTERN RECOGNITION AND MACHINE LEARNING PATTERN RECOGNITION AND MACHINE LEARNING Chapter 1. Introduction Shuai Huang April 21, 2014 Outline 1 What is Machine Learning? 2 Curve Fitting 3 Probability Theory 4 Model Selection 5 The curse of dimensionality

More information

Heriot-Watt University

Heriot-Watt University Heriot-Watt University Heriot-Watt University Research Gateway Prediction of settlement delay in critical illness insurance claims by using the generalized beta of the second kind distribution Dodd, Erengul;

More information

Some Curiosities Arising in Objective Bayesian Analysis

Some Curiosities Arising in Objective Bayesian Analysis . Some Curiosities Arising in Objective Bayesian Analysis Jim Berger Duke University Statistical and Applied Mathematical Institute Yale University May 15, 2009 1 Three vignettes related to John s work

More information

STA414/2104. Lecture 11: Gaussian Processes. Department of Statistics

STA414/2104. Lecture 11: Gaussian Processes. Department of Statistics STA414/2104 Lecture 11: Gaussian Processes Department of Statistics www.utstat.utoronto.ca Delivered by Mark Ebden with thanks to Russ Salakhutdinov Outline Gaussian Processes Exam review Course evaluations

More information

Quantitative Interpretation

Quantitative Interpretation Quantitative Interpretation The aim of quantitative interpretation (QI) is, through the use of amplitude analysis, to predict lithology and fluid content away from the well bore. This process should make

More information

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j Standard Errors & Confidence Intervals β β asy N(0, I( β) 1 ), where I( β) = [ 2 l(β, φ; y) ] β i β β= β j We can obtain asymptotic 100(1 α)% confidence intervals for β j using: β j ± Z 1 α/2 se( β j )

More information

1 Hypothesis Testing and Model Selection

1 Hypothesis Testing and Model Selection A Short Course on Bayesian Inference (based on An Introduction to Bayesian Analysis: Theory and Methods by Ghosh, Delampady and Samanta) Module 6: From Chapter 6 of GDS 1 Hypothesis Testing and Model Selection

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

The connection of dropout and Bayesian statistics

The connection of dropout and Bayesian statistics The connection of dropout and Bayesian statistics Interpretation of dropout as approximate Bayesian modelling of NN http://mlg.eng.cam.ac.uk/yarin/thesis/thesis.pdf Dropout Geoffrey Hinton Google, University

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Bayesian Inference: Concept and Practice

Bayesian Inference: Concept and Practice Inference: Concept and Practice fundamentals Johan A. Elkink School of Politics & International Relations University College Dublin 5 June 2017 1 2 3 Bayes theorem In order to estimate the parameters of

More information

Introduction to Maximum Likelihood Estimation

Introduction to Maximum Likelihood Estimation Introduction to Maximum Likelihood Estimation Eric Zivot July 26, 2012 The Likelihood Function Let 1 be an iid sample with pdf ( ; ) where is a ( 1) vector of parameters that characterize ( ; ) Example:

More information

Bayesian rules of probability as principles of logic [Cox] Notation: pr(x I) is the probability (or pdf) of x being true given information I

Bayesian rules of probability as principles of logic [Cox] Notation: pr(x I) is the probability (or pdf) of x being true given information I Bayesian rules of probability as principles of logic [Cox] Notation: pr(x I) is the probability (or pdf) of x being true given information I 1 Sum rule: If set {x i } is exhaustive and exclusive, pr(x

More information

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci

More information

Behavioral Data Mining. Lecture 7 Linear and Logistic Regression

Behavioral Data Mining. Lecture 7 Linear and Logistic Regression Behavioral Data Mining Lecture 7 Linear and Logistic Regression Outline Linear Regression Regularization Logistic Regression Stochastic Gradient Fast Stochastic Methods Performance tips Linear Regression

More information

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE Data Provided: None DEPARTMENT OF COMPUTER SCIENCE Autumn Semester 203 204 MACHINE LEARNING AND ADAPTIVE INTELLIGENCE 2 hours Answer THREE of the four questions. All questions carry equal weight. Figures

More information

Nonparmeteric Bayes & Gaussian Processes. Baback Moghaddam Machine Learning Group

Nonparmeteric Bayes & Gaussian Processes. Baback Moghaddam Machine Learning Group Nonparmeteric Bayes & Gaussian Processes Baback Moghaddam baback@jpl.nasa.gov Machine Learning Group Outline Bayesian Inference Hierarchical Models Model Selection Parametric vs. Nonparametric Gaussian

More information

L 0 methods. H.J. Kappen Donders Institute for Neuroscience Radboud University, Nijmegen, the Netherlands. December 5, 2011.

L 0 methods. H.J. Kappen Donders Institute for Neuroscience Radboud University, Nijmegen, the Netherlands. December 5, 2011. L methods H.J. Kappen Donders Institute for Neuroscience Radboud University, Nijmegen, the Netherlands December 5, 2 Bert Kappen Outline George McCullochs model The Variational Garrote Bert Kappen L methods

More information

Bayesian model selection: methodology, computation and applications

Bayesian model selection: methodology, computation and applications Bayesian model selection: methodology, computation and applications David Nott Department of Statistics and Applied Probability National University of Singapore Statistical Genomics Summer School Program

More information

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem

More information

Transdimensional Markov Chain Monte Carlo Methods. Jesse Kolb, Vedran Lekić (Univ. of MD) Supervisor: Kris Innanen

Transdimensional Markov Chain Monte Carlo Methods. Jesse Kolb, Vedran Lekić (Univ. of MD) Supervisor: Kris Innanen Transdimensional Markov Chain Monte Carlo Methods Jesse Kolb, Vedran Lekić (Univ. of MD) Supervisor: Kris Innanen Motivation for Different Inversion Technique Inversion techniques typically provide a single

More information

Variational Methods in Bayesian Deconvolution

Variational Methods in Bayesian Deconvolution PHYSTAT, SLAC, Stanford, California, September 8-, Variational Methods in Bayesian Deconvolution K. Zarb Adami Cavendish Laboratory, University of Cambridge, UK This paper gives an introduction to the

More information

CMU-Q Lecture 24:

CMU-Q Lecture 24: CMU-Q 15-381 Lecture 24: Supervised Learning 2 Teacher: Gianni A. Di Caro SUPERVISED LEARNING Hypotheses space Hypothesis function Labeled Given Errors Performance criteria Given a collection of input

More information

Bayesian inference J. Daunizeau

Bayesian inference J. Daunizeau Bayesian inference J. Daunizeau Brain and Spine Institute, Paris, France Wellcome Trust Centre for Neuroimaging, London, UK Overview of the talk 1 Probabilistic modelling and representation of uncertainty

More information

Bayesian Learning (II)

Bayesian Learning (II) Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning (II) Niels Landwehr Overview Probabilities, expected values, variance Basic concepts of Bayesian learning MAP

More information

Making rating curves - the Bayesian approach

Making rating curves - the Bayesian approach Making rating curves - the Bayesian approach Rating curves what is wanted? A best estimate of the relationship between stage and discharge at a given place in a river. The relationship should be on the

More information

Net-to-gross from Seismic P and S Impedances: Estimation and Uncertainty Analysis using Bayesian Statistics

Net-to-gross from Seismic P and S Impedances: Estimation and Uncertainty Analysis using Bayesian Statistics Net-to-gross from Seismic P and S Impedances: Estimation and Uncertainty Analysis using Bayesian Statistics Summary Madhumita Sengupta*, Ran Bachrach, Niranjan Banik, esterngeco. Net-to-gross (N/G ) is

More information

Bayesian Inference for Normal Mean

Bayesian Inference for Normal Mean Al Nosedal. University of Toronto. November 18, 2015 Likelihood of Single Observation The conditional observation distribution of y µ is Normal with mean µ and variance σ 2, which is known. Its density

More information

CSC2515 Assignment #2

CSC2515 Assignment #2 CSC2515 Assignment #2 Due: Nov.4, 2pm at the START of class Worth: 18% Late assignments not accepted. 1 Pseudo-Bayesian Linear Regression (3%) In this question you will dabble in Bayesian statistics and

More information

Reservoir connectivity uncertainty from stochastic seismic inversion Rémi Moyen* and Philippe M. Doyen (CGGVeritas)

Reservoir connectivity uncertainty from stochastic seismic inversion Rémi Moyen* and Philippe M. Doyen (CGGVeritas) Rémi Moyen* and Philippe M. Doyen (CGGVeritas) Summary Static reservoir connectivity analysis is sometimes based on 3D facies or geobody models defined by combining well data and inverted seismic impedances.

More information

Bayesian Model Comparison

Bayesian Model Comparison BS2 Statistical Inference, Lecture 11, Hilary Term 2009 February 26, 2009 Basic result An accurate approximation Asymptotic posterior distribution An integral of form I = b a e λg(y) h(y) dy where h(y)

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information

p(z)

p(z) Chapter Statistics. Introduction This lecture is a quick review of basic statistical concepts; probabilities, mean, variance, covariance, correlation, linear regression, probability density functions and

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Multivariate statistical methods and data mining in particle physics

Multivariate statistical methods and data mining in particle physics Multivariate statistical methods and data mining in particle physics RHUL Physics www.pp.rhul.ac.uk/~cowan Academic Training Lectures CERN 16 19 June, 2008 1 Outline Statement of the problem Some general

More information

Statistical Machine Learning Hilary Term 2018

Statistical Machine Learning Hilary Term 2018 Statistical Machine Learning Hilary Term 2018 Pier Francesco Palamara Department of Statistics University of Oxford Slide credits and other course material can be found at: http://www.stats.ox.ac.uk/~palamara/sml18.html

More information

Data-Driven Bayesian Model Selection: Parameter Space Dimension Reduction using Automatic Relevance Determination Priors

Data-Driven Bayesian Model Selection: Parameter Space Dimension Reduction using Automatic Relevance Determination Priors Data-Driven : Parameter Space Dimension Reduction using Priors Mohammad Khalil mkhalil@sandia.gov, Livermore, CA Workshop on Uncertainty Quantification and Data-Driven Modeling Austin, Texas March 23-24,

More information

Wellcome Trust Centre for Neuroimaging, UCL, UK.

Wellcome Trust Centre for Neuroimaging, UCL, UK. Bayesian Inference Will Penny Wellcome Trust Centre for Neuroimaging, UCL, UK. SPM Course, Virginia Tech, January 2012 What is Bayesian Inference? (From Daniel Wolpert) Bayesian segmentation and normalisation

More information

QUANTITATIVE INTERPRETATION

QUANTITATIVE INTERPRETATION QUANTITATIVE INTERPRETATION THE AIM OF QUANTITATIVE INTERPRETATION (QI) IS, THROUGH THE USE OF AMPLITUDE ANALYSIS, TO PREDICT LITHOLOGY AND FLUID CONTENT AWAY FROM THE WELL BORE This process should make

More information

Edinburgh Anisotropy Project, British Geological Survey, Murchison House, West Mains

Edinburgh Anisotropy Project, British Geological Survey, Murchison House, West Mains Frequency-dependent AVO attribute: theory and example Xiaoyang Wu, 1* Mark Chapman 1,2 and Xiang-Yang Li 1 1 Edinburgh Anisotropy Project, British Geological Survey, Murchison House, West Mains Road, Edinburgh

More information

Modelling behavioural data

Modelling behavioural data Modelling behavioural data Quentin Huys MA PhD MBBS MBPsS Translational Neuromodeling Unit, ETH Zürich Psychiatrische Universitätsklinik Zürich Outline An example task Why build models? What is a model

More information

Conditional Random Field

Conditional Random Field Introduction Linear-Chain General Specific Implementations Conclusions Corso di Elaborazione del Linguaggio Naturale Pisa, May, 2011 Introduction Linear-Chain General Specific Implementations Conclusions

More information

Multivariate Bayesian Linear Regression MLAI Lecture 11

Multivariate Bayesian Linear Regression MLAI Lecture 11 Multivariate Bayesian Linear Regression MLAI Lecture 11 Neil D. Lawrence Department of Computer Science Sheffield University 21st October 2012 Outline Univariate Bayesian Linear Regression Multivariate

More information

An Extended BIC for Model Selection

An Extended BIC for Model Selection An Extended BIC for Model Selection at the JSM meeting 2007 - Salt Lake City Surajit Ray Boston University (Dept of Mathematics and Statistics) Joint work with James Berger, Duke University; Susie Bayarri,

More information

A Frequentist Assessment of Bayesian Inclusion Probabilities

A Frequentist Assessment of Bayesian Inclusion Probabilities A Frequentist Assessment of Bayesian Inclusion Probabilities Department of Statistical Sciences and Operations Research October 13, 2008 Outline 1 Quantitative Traits Genetic Map The model 2 Bayesian Model

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

An introduction to Bayesian inference and model comparison J. Daunizeau

An introduction to Bayesian inference and model comparison J. Daunizeau An introduction to Bayesian inference and model comparison J. Daunizeau ICM, Paris, France TNU, Zurich, Switzerland Overview of the talk An introduction to probabilistic modelling Bayesian model comparison

More information

Model Choice. Hoff Chapter 9, Clyde & George Model Uncertainty StatSci, Hoeting et al BMA StatSci. October 27, 2015

Model Choice. Hoff Chapter 9, Clyde & George Model Uncertainty StatSci, Hoeting et al BMA StatSci. October 27, 2015 Model Choice Hoff Chapter 9, Clyde & George Model Uncertainty StatSci, Hoeting et al BMA StatSci October 27, 2015 Topics Variable Selection / Model Choice Stepwise Methods Model Selection Criteria Model

More information