Calibration and uncertainty quantification using multivariate simulator output

Size: px
Start display at page:

Download "Calibration and uncertainty quantification using multivariate simulator output"

Transcription

1 Calibration and uncertainty quantification using multivariate simulator output LA-UR 4-74 Dave Higdon, Statistical Sciences Group, LANL Jim Gattiker, Statistical Sciences Group, LANL Brian Williams, Statistical Sciences Group, LANL

2 Inference combining a physics model with experimental data drop time drop height (floor) Data generated from model: d z =.3 dz dt dt + ǫ simulation model: d z = dt drop time statistical model: y(z) = η(z) + δ(z) + ǫ drop height (floor) drop time Improved physics model: d z = θ dz dt dt + ǫ statistical model: y(z) = η(z, θ) + δ(z) + ǫ drop height (floor)

3 Basic formulation borrows from Kennedy and O Hagan () 7 (a) (b) (c) (d) η(x), y(x). η(x), y(x). δ(x). y(x) x 3. x x x experimental conditions θ calibration parameters ζ(x) true physical system response given inputs x η(x, θ) simulator response at x and θ. y(x) experimental observation of the physical system δ(x) discrepancy between ζ(x) and η(x, θ) may be decomposed into numerical error and bias e(x) observation error of the experimental data y(x) = ζ(x) + e(x) y(x) = η(x, θ) + δ(x) + e(x) 3. x

4 A Bayesian approach for combining simulations and experimental data for forecasting, calibration and uncertainty quantification prior uncertainty posterior uncertainty 3 3 y(x), η(x,θ) y(x), η(x,θ). θ x. θ x A simple example... x model or system inputs θ model calibration parameters ζ(x) true physical system response given inputs x η(x, θ) simulator response at x and θ. y(x) experimental observation of the physical system e(x) observation error of the experimental data Assume: y(x) = ζ(x) + e(x) = η(x, θ) + e(x) θ unknown. Standard Bayesian estimation gives: π(θ y(x)) L(y(x) η(x, θ)) π(θ)

5 Accounting for limited simulator runs y(x), η(x,θ) 3 prior uncertainty. θ 3. x Borrows from Kennedy and O Hagan (). x model or system inputs θ calibration parameters ζ(x) true physical system response given inputs x η(x, θ) simulator response at x and θ. simulator run at limited input settings η = (η(x, θ),..., η(x m, θm)) T treat η(, ) as a random function use GP prior for η(, ) y(x) experimental observation of the physical system e(x) observation error of the experimental data y(x) = ζ(x) + e(x) y(x) = η(x, θ) + e(x)

6 Accounting for limited simulation runs prior uncertainty posterior realizations of η(x,t) posterior uncertainty 3 3 y(x), η(x,θ). θ 3. x Again, standard Bayesian estimation gives: η(x,t). x π(θ, η(, ) y(x)) L(y(x) η(x, θ)) π(θ) π(η(, )) t. y(x), η(x,θ). θ 3. x Posterior means and quantiles shown. Uncertainty in θ and η(x, θ) are incorporated into the forecast. Gaussian process models for η(, ).

7 Accounting for model discrepancy prior uncertainty Borrows from Kennedy and O Hagan (). y(x), η(x,θ) 3. θ 3. x x model or system inputs θ model or system inputs ζ(x) true physical system response given inputs x η(x, θ) simulator response at x and θ. y(x) experimental observation of the physical system δ(x) discrepancy between ζ(x) and η(x, θ) may be decomposed into numerical error and bias e(x) observation error of the experimental data y(x) = ζ(x) + e(x) y(x) = η(x, θ) + δ(x) + e(x) y(x) = η(x, θ) + δ n (x) + δ b (x) + e(x)

8 Accounting for model discrepancy prior uncertainty posterior model uncertainty y(x), η(x,θ) δ(x) 3. θ 3. x 3 posterior model discrepancy 3. x y(x), η(x,θ) y(x) 3. θ 3. x 3 calibrated forecast 3. x Again, standard Bayesian estimation gives: π(θ, δ n, δ b y(x)) L(y(x) η(x, θ), δ(x)) π(θ) π(η) π(δ) Posterior means and 9% CI s shown. Posterior prediction for ζ(x) is obtained by computing the posterior distribution for η(x, θ) + δ(x) Uncertainty in θ, η(x, t), and δ(x) are incorporated into the forecast. Gaussian process models for η(x, t) and δ(x)

9 Application: implosions of steel cylinders Neddermeyer 43 Initial work on implosion for fat man. Use high explosive (HE) to crush steel cylindrical shells Investigate the feasability of a controlled implosion

10 Some History Early work on cylinders called beer can experiments. Early work not encouraging:...i question Dr. Neddermeyer s seriousness... Deke Parsons. It stinks. R. Feynman Teller and VonNeumann were quite supportive of the implosion idea Data on collapsing cylinder from high speed photography. Symmetrical implosion eventually accomplished using HE lenses by Kistiakowsky. Implosion played a key role in early computer experiments. Feynman worked on implosion calculations with IBM accounting machines. Eventually first computer with addressable memory was developed (MANIAC ).

11 The Experiments

12 Neddermeyer s Model cm 3 4 X s 3 4 X s 3 4 X s 3 4 X s 3 4 X s cm 3 4 X s 3 4 X s 3 4 X s 3 4 X s 3 4 X s Energy from HE imparts an initial inward velocity to the cylinder v = m e u m + m e /m mass ratio m e /m of HE to steel; u energy per unit mass from HE. Energy converts to work done on the cylinder: s { work per unit mass = w = r ρ( λ) i log ri ro log ro + λ log λ } r i = scaled inner radius; r o = scaled outer radius; λ = initial r i /r o ; s = steel yielding stress; ρ = density of steel.

13 Neddermeyer s Model cm 3 4 X s 3 4 X s 3 4 X s 3 4 X s 3 4 X s where cm 3 4 X s 3 4 X s 3 4 X s 3 4 X s 3 4 X s ODE: dr dt = [ R f(r) { v s ρ g(r) }] r = inner radius of cylinder varies with time R = initial outer ( radius of cylinder f(r) = r r λ ln + λ ) r g(r) = ( λ ) [r lnr (r + λ ) ln(r + λ ) λ lnλ ] λ = initial ratio of cylinder r(t = )/R constant volume condition: r outer r = λ

14 Goal: use experimental data to calibrate s and u ; obtain prediction uncertainty for new experiment expt cm expt 3 t = µs cm expt t = µs cm t = 4 µs m e /m.3 m e /m.7 m e /m.36 Hypothetical data obtained from photos at different times during the 3 experimental implosions. All cylinders had a.in outer and a.in inner radius. (λ = 3 ).

15 Carry out simulated implosions using Neddermeyer s model Sequence of runs carried at m input settings (x, θ, θ) = (m e /m, s, u ) varying x over predefined ranges using an OA(3, 4 3 θ θ )-based LH design.... x m θm θm.. inner radius (cm). inner radius (cm) time (s) x 3 x 4 time (s) pi angle (radians) pi radius by time radius by time and angle φ. Each simulation produces a n η = 6 vector of radii for times 6 angles.

16 Generating OA-based LH designs Example: N = 6, 3 factors each at 4 levels OA(6, 4 3 ) design induced LH design x_ x x_ x_ x x Ensures some higher dimensional filling relative to standard LH designs.

17 Generating (nearly) OA-based LH designs Example: NOA(48, 4 8 ) N = 48, 8 factors each at 4 levels..4.8 x_ x_ x_ columns of NOA design matrix X are not exactly orthogonal allows more factors with good higher dimensional properties x_4 x_ x_6 x_ x_

18 PC representation of simulation output Ξ = [η ; ;η m ] a n η m matrix that holds output of m simulations SVD decomposition: Ξ = UDV T K η is st p η columns of [ m UD] columns of [ mv T ] have variance Cylinder example: PC (9.4% of variation) PC (.9% of variation) PC 3 (.7% of variation) r.. time angle r.. time angle r.. time angle p η = 3 PC s: K η = [k ;k ;k 3 ] each vector k i holds trace of PC i. k i s do not change with φ from symmetry of Neddermeyer s model. Simulated trace η(x i, θ i, θ i ) = K ηw(x i, θ i, θ i )+ǫ i, ǫ i s iid N(, λ η ), for any set of tried simulation inputs (x i, θ i, θ i ).

19 Gaussian process models for PC weights Want to evaluate η(x, θ, θ ) at arbitrary input setting (x, θ, θ ). Also want analysis to account for uncertainty here. Approach: model each PC weight as a Gaussian process: where w i (x, θ, θ ) GP(, λ wi R((x, θ), (x, θ ); ρ wi )) R((x, θ), (x, θ ); ρ wi ) = Restricting to the design settings p x k= ρ 4(x k x k ) wik p θ k= x θ θ... and specifying x m θm θm w i = (w i (x, θ, θ ),..., w i (x m, θ m, θ m)) T gives iid w i N (, λ wi R((x, θ ); ρ wi ) ), i =,..., p η where R((x, θ ); ρ wi ) m m is given by (). ρ 4(θ k θ k ) wi(k+p x ) () *note: additional nugget term w i iid N (, λ wi R((x, θ ); ρ wi ) + λ ǫi I m), i =,...,pη, may be useful.

20 Gaussian process models for PC weights At the m simulation input settings the mp η -vector w has prior disribution w λ w R((x, θ ); ρ w ) w =. N.,... w pη λ wp η R((x, θ ); ρ wpη ) w N(, Σ w ); note Σ w = I pη λ w R((x, θ ); ρ w ) can break down. Emulator likelihood: η = vec([η(x, θ, θ ); ;η(x m, θ m, θ m)]) mnη L(η w, λ η ) λη exp { λ η(η Kw) T (η Kw) }, λ η Γ(a η, b η ) where n η is the number of observations in a simulated trace and Equivalently K = [I m k ; ;I m k pη ]. L(η w, λ η ) λ mp η η exp { λ η(w ŵ) T (K T K)(w ŵ) } m(nη pη) λη exp { λ ηη T (I K(K T K) K T )η } mpη λη exp { λ η(w ŵ) T (K T K)(w ŵ) }, λ η Γ(a η, b η) a η = a η + m(n η p η ), b η = b η + ηt (I K(K T K) K T )η, ŵ = (K T K) K T η.

21 Gaussian process models for PC weights Resulting posterior can then be based on computed PC weights ŵ: ŵ w,λ η N(w, (λ η K T K) ) w λ w, ρ w N(, Σ w ) ŵ λ η, λ w, ρ w N(, (λ η K T K) + Σ w ) Resulting posterior is then: π(λ η, λ w, ρ w ŵ) (λη K T K) + Σ w exp{ ŵt ([λ η K T K] + Σ w ) ŵ} p η λ a η η e b ηλ η λ a w wi e b wλ wi p η i= p x j= i= ( ρ wij ) b ρ p θ ( ρ wi(j+px )) b ρ j= MCMC via Metropolis works fine here. Bounded range of ρ wij s facilitates MCMC.

22 Posterior distribution of ρ w.8 PC [x θ] PC. 3 [x θ] PC3. 3 [x θ] Separate models by PC More opportunity to take advantage of effect sparsity

23 Predicting simulator output at untried (x, θ, θ ) Want η(x, θ, θ ) = Kw(x, θ, θ ) For a given draw (λ η, λ w, ρ w ) a draw of w can be produced: ( ) (( ) [( ) ]) ŵ (λη K w N, T K) + Σ w,w (λ w, ρ w ) Define Then ( ) V V V = = V V [( ) (λη K T K) ] + Σ w,w (λ w, ρ w ) w ŵ N(V V ŵ, V V V V ) Realizations can be generated from sample of MCMC output. Lots of info (data?) makes conditioning on point estimate ( λ η, λ w, ρ w ) a good approximation to the posterior. Posterior mean or median work well for ( λ η, λ w, ρ w )

24 Exploring sensitivity of simulator output to model inputs Simulator predictions varing input, holding others at nominal

25 Basic formulation borrows from Kennedy and O Hagan () Experiment r, η δ x 4 time x 4 time r, η+δ x 4 time φ φ φ δ x = m e /m.3 θ = s? θ = u? (t, φ) simulation output space x experimental conditions θ calibration parameters ζ(x) true physical system response given conditions x η(x, θ) simulator response at x and θ. y(x) experimental observation of the physical system δ(x) discrepancy between ζ(x) and η(x, θ) may be decomposed into numerical error and bias e(x) observation error of the experimental data y(x) = ζ(x) + e(x) y(x) = η(x, θ) + δ(x) + e(x)

26 Kernel basis representation for spatial processes δ(s) Define p δ basis functions d (s),..., d pδ (s). basis Here d j (s) is normal density cetered at spatial location ω j : d j (s) = π exp{ (s ω j) } set δ(s) = p δ j= s d j (s)v j where v N(, λ v I pδ ). Can represent δ = (δ(s ),..., δ(s n )) T as δ = Dv where D ij = d j (s i )

27 v and d(s) determine spatial processes δ(s) d j (s)v j δ(s) basis... z(s) s s Continuous representation: δ(s) = p δ j= d j (s)v j where v N(, λ v I pδ ). Discrete representation: For δ = (δ(s ),..., δ(s n )) T, δ = Dv where D ij = d j (s i )

28 Basis representation of discrepancy time angle φ Represent discrepancy δ(x) using basis functions and weights p δ = 4 basis functions over (t,φ); D = [d ; ;d pδ ]; d k s hold basis. δ(x) = Dv(x) where v(x) GP (, λ v I pδ R(x, x ;ρ v ) ) with R(x, x ;ρ v ) = p x k= ρ 4(x k x k ) vk ()

29 Integrated model formulation Data y(x ),..., y(x n ) collected for n experiments at input conditions x,..., x n. Each y(x i ) is a collection of n yi measurements over points indexed by (t, φ). y(x i ) = η(x i, θ) + δ(x i ) + e i = K i ( w(x i, θ) + ( D i v(x i ) + ) e i ) v(xi ) y(x i ) w(x i, θ), v(x i ), λ y N [D i ;K i ], (λ w(x i, θ) y W i ) Since support of each y(x i ) varies and doesn t match that of sims, the basis vectors in K i must be interpolated from K η ; similary, D i must be computed from the support of y(x i ): x r. r.. r x time pi pi angle pi angle pi time 4 x pi pi angle time x *note: cubic spline interpolation over (time, φ) used here.

30 Define Integrated model formulation n y = n y + + n yn, the total number of experimental data points, y to be the n y -vector from concatination of the y(x i ) s, v = vec([v(x ); ;v(x n )] T ) and u(θ) = vec([w(x, θ, θ ); ;w(x n, θ, θ )] T ) ( ( ) ) v y v, u(θ), λ y N B, (λ u(θ) y W y ), λ y Γ(a y, b y ) (3) where W y = diag(w,..., W n ) and ( ) P T B = diag(d,..., D n, K,..., K n ) D PK T P D and P K are permutation matricies whose rows are given by: P D (j + n(i ); ) = e T (j )p δ +i, i =,..., p δ; j =,..., n P K (j + n(i ); ) = e T (j )p η +i, i =,..., p η; j =,..., n

31 Integrated model formulation (continued) Equivalently (3) can be represented ) ( ) (( ) ) (ˆv v v, λ û u(θ) y N, (λ u(θ) y B T W y B), λ y Γ(a y, b y) with n y = n y + + n yn, the total number of experimental data points ) (ˆv = (B T W û y B) B T W y y a y = a y + [n y n(p δ + p η )] [ ( b y = b y + )) T ( )) (ˆv (ˆv ] y B W û y y B û model dimension reduction simulator data and discrep standard n η m n y basis p η m n (p δ + p η ) Basis approach particularly efficient when n η and n y are large.

32 λ Marginal likelihood The (marginal) likelihood L(ˆv, û, ŵ λ η, λ w, ρ w, λ y, λ v, ρ v, θ) has the form ˆv Σ v û N, Λ y + ŵ Λ Σ η uw where Λ y = λ y B T W y B Λ η = λ η K T K Σ v = λ v I pη R(x, x;ρ v ) R(x, x;ρ v ) = n n correlation matrix from applying () to the conditions x,..., x n corresponding the the n experiments. Σ uw = w R((x, θ), (x, θ); ρ w) λ w R((x, θ), (x, θ ); ρ w ) λ wp η R((x, θ), (x, θ); ρ wpη ) λ wp η R((x, θ), (x, θ ); ρ wpη ) w R((x, θ ), (x, θ); ρ w ) λ w R((x, θ ), (x, θ ); ρ w ) λ wp η R((x, θ ), (x, θ); ρ wpη ) λ wp η R((x, θ ), (x, θ ); ρ wpη ) λ Permutation of Σ uw is block diagonal can speed up computations. Only off diagonal blocks of Σ uw depend on θ.

33 Posterior distribution Likelihood: L(ˆv, û, ŵ λ η, λ w, ρ w, λ y, λ v, ρ v, θ) Prior: π(λ η, λ w, ρ w, λ y, λ v, ρ v, θ) Posterior: π(λ η, λ w, ρ w, λ y, λ v, ρ v, θ ˆv, û, ŵ) L(ˆv, û, ŵ λ η, λ w, ρ w, λ y, λ v, ρ v, θ) π(λ η, λ w, ρ w, λ y, λ v, ρ v, θ) Posterior exploration via MCMC Can take advantage of structure and sparcity to speed up sampling. A useful approximation to speed up posterior evaluation: π(λ η, λ w, ρ w, λ y, λ v, ρ v, θ ˆv, û, ŵ) L(ŵ λ η, λ w, ρ w ) π(λ η, λ w, ρ w ) L(ˆv, û λ η, λ w, ρ w, λ y, λ v, ρ v, θ) π(λ y, λ v, ρ v, θ) In this approximation, experimental data is not used to inform about parameters λ η, λ w, ρ w which govern the simulator process η(x, θ).

34 Posterior distribution of model parameters (θ, θ )

35 Posterior mean decomposition for each experiment Experiment Experiment Experiment 3 r, η r, η r, η x 4 time φ x 4 time φ x 4 time φ δ δ δ x 4 time φ x 4 time φ x 4 time φ r, η+δ r, η+δ r, η+δ x 4 time φ x 4 time φ x 4 time φ

36 Posterior prediction for implosion in each experiment Experiment Experiment Experiment 3 r... x time φ r... x time φ r... x time φ Experiment r... x time φ r

37 9% prediction intervals for implosions at exposure times Experiment Experiment Experiment 3 Time µs Time 7 µs Time µs Time 4 µs Time 4 µs Time 4 µs Predictions from separate analyses which hold data from the experiment being predicted.

38 Quantifying uncertainty for simulation-based forecasts Simulation-based predictions accumulate uncertainty due to: parameter/calibration uncertainty; simulator discrepancy/inadequacy; observation error in data; sparseness in data Limits on dimensionality? Have dealt with up to -dimensional θ. Statistics typically uses the wrong model (eg. a regression line) to explain data. So the framework is nothing new, in principle. Calibrating a model with substantial inadequacy? only slight inadequacy? The slowness of the simulator and high dimensionality make things difficult. Extrapolation is often a goal in such investigations. Generally, the the closer to reality the simulator is, the more it can be trusted for extrapolation. Can this be more rigorously formalized?

39 Application: Ta Flyerplate Experiment PTW model governs features on the visar velocity profile. Use pricipal components (EOF s) to deal with high dimensionality.

40 Simulations and Data.3 velocity km/s..... standardized velocity.. 3 time µs.. 3 time µs PTW calibration parameters with input domains Parameter Description Domain Min Max θ Initial strain hardening rate Material constant in thermal activation κ energy term relates to the temperature dependence Material constant in thermal activation log(γ) energy term relates to the strain rate dependence y Maximum yield stress (at K) y Minimum yield stress ( melting)..8 s Maximum saturation stress (at K).4.36 s Minimum saturation stress ( melting).8.43

41 PC representation of simulation output.. standardized velocity time (µ s) fit using p η = 7% of variation velocity (km/s) time (µ s) standardized velocity.. 3 time (µ s)

42 Marginal Posteriors for spatial correlation parameters ρ wij PC.9.9 PC PC θ θ θ

43 PC - based sensitivities

44 .8 Discrepancy basis stadardized velocity time (µ s) Represent discrepancy δ(x) using basis functions and weights Here d j (s) is normal density cetered at spatial location ω j : d j (s) = exp{ π (s ω j) } set δ(s) = p δ j= d j (s)v j where v N(, λ v I pδ ). Can represent δ = (δ(s ),..., δ(s n )) T as δ = Dv where p δ = basis functions over t. D ij = d j (s i )

45 posterior distribution for PTW parameters

46 posterior predictive distribution for trace η Y time (µs) time (µs) standardized velocity standardized velocity 4 δ time (µs) time (µs) standardized velocity standardized velocity

47 posterior predictive distribution for trace (original scale) η Y velocity. velocity time (µs)... 3 time (µs).3 δ.3.. velocity. velocity time (µs)... 3 time (µs)

Computer Model Calibration using High Dimensional Output

Computer Model Calibration using High Dimensional Output Computer Model Calibration using High Dimensional Output Dave Higdon, Los Alamos National Laboratory Jim Gattiker, Los Alamos National Laboratory Brian Williams, Los Alamos National Laboratory Maria Rightley,

More information

Combining Experimental Data and Computer Simulations, With an Application to Flyer Plate Experiments

Combining Experimental Data and Computer Simulations, With an Application to Flyer Plate Experiments Combining Experimental Data and Computer Simulations, With an Application to Flyer Plate Experiments LA-UR-6- Brian Williams, Los Alamos National Laboratory Dave Higdon, Los Alamos National Laboratory

More information

Sequential Importance Sampling for Rare Event Estimation with Computer Experiments

Sequential Importance Sampling for Rare Event Estimation with Computer Experiments Sequential Importance Sampling for Rare Event Estimation with Computer Experiments Brian Williams and Rick Picard LA-UR-12-22467 Statistical Sciences Group, Los Alamos National Laboratory Abstract Importance

More information

Uncertainty quantification and calibration of computer models. February 5th, 2014

Uncertainty quantification and calibration of computer models. February 5th, 2014 Uncertainty quantification and calibration of computer models February 5th, 2014 Physical model Physical model is defined by a set of differential equations. Now, they are usually described by computer

More information

Bayesian inference & process convolution models Dave Higdon, Statistical Sciences Group, LANL

Bayesian inference & process convolution models Dave Higdon, Statistical Sciences Group, LANL 1 Bayesian inference & process convolution models Dave Higdon, Statistical Sciences Group, LANL 2 MOVING AVERAGE SPATIAL MODELS Kernel basis representation for spatial processes z(s) Define m basis functions

More information

An introduction to Bayesian statistics and model calibration and a host of related topics

An introduction to Bayesian statistics and model calibration and a host of related topics An introduction to Bayesian statistics and model calibration and a host of related topics Derek Bingham Statistics and Actuarial Science Simon Fraser University Cast of thousands have participated in the

More information

Gaussian Process Modeling in Cosmology: The Coyote Universe. Katrin Heitmann, ISR-1, LANL

Gaussian Process Modeling in Cosmology: The Coyote Universe. Katrin Heitmann, ISR-1, LANL 1998 2015 Gaussian Process Modeling in Cosmology: The Coyote Universe Katrin Heitmann, ISR-1, LANL In collaboration with: Jim Ahrens, Salman Habib, David Higdon, Chung Hsu, Earl Lawrence, Charlie Nakhleh,

More information

Fast Dimension-Reduced Climate Model Calibration and the Effect of Data Aggregation

Fast Dimension-Reduced Climate Model Calibration and the Effect of Data Aggregation Fast Dimension-Reduced Climate Model Calibration and the Effect of Data Aggregation Won Chang Post Doctoral Scholar, Department of Statistics, University of Chicago Oct 15, 2014 Thesis Advisors: Murali

More information

Nonlinear Models. and. Hierarchical Nonlinear Models

Nonlinear Models. and. Hierarchical Nonlinear Models Nonlinear Models and Hierarchical Nonlinear Models Start Simple Progressively Add Complexity Tree Allometries Diameter vs Height with a hierarchical species effect Three response variables: Ht, crown depth,

More information

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the

More information

Bayesian Dynamic Linear Modelling for. Complex Computer Models

Bayesian Dynamic Linear Modelling for. Complex Computer Models Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer

More information

SURROGATE PREPOSTERIOR ANALYSES FOR PREDICTING AND ENHANCING IDENTIFIABILITY IN MODEL CALIBRATION

SURROGATE PREPOSTERIOR ANALYSES FOR PREDICTING AND ENHANCING IDENTIFIABILITY IN MODEL CALIBRATION International Journal for Uncertainty Quantification, ():xxx xxx, 0 SURROGATE PREPOSTERIOR ANALYSES FOR PREDICTING AND ENHANCING IDENTIFIABILITY IN MODEL CALIBRATION Zhen Jiang, Daniel W. Apley, & Wei

More information

1. Gaussian process emulator for principal components

1. Gaussian process emulator for principal components Supplement of Geosci. Model Dev., 7, 1933 1943, 2014 http://www.geosci-model-dev.net/7/1933/2014/ doi:10.5194/gmd-7-1933-2014-supplement Author(s) 2014. CC Attribution 3.0 License. Supplement of Probabilistic

More information

Gaussian processes for uncertainty quantification in computer experiments

Gaussian processes for uncertainty quantification in computer experiments Gaussian processes for uncertainty quantification in computer experiments Richard Wilkinson University of Nottingham Gaussian process summer school, Sheffield 2013 Richard Wilkinson (Nottingham) GPs for

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Construction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting

Construction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting Construction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting Anne Philippe Laboratoire de Mathématiques Jean Leray Université de Nantes Workshop EDF-INRIA,

More information

Hierarchical Modeling for Univariate Spatial Data

Hierarchical Modeling for Univariate Spatial Data Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This

More information

Gaussian with mean ( µ ) and standard deviation ( σ)

Gaussian with mean ( µ ) and standard deviation ( σ) Slide from Pieter Abbeel Gaussian with mean ( µ ) and standard deviation ( σ) 10/6/16 CSE-571: Robotics X ~ N( µ, σ ) Y ~ N( aµ + b, a σ ) Y = ax + b + + + + 1 1 1 1 1 1 1 1 1 1, ~ ) ( ) ( ), ( ~ ), (

More information

Hierarchical Modelling for Univariate and Multivariate Spatial Data

Hierarchical Modelling for Univariate and Multivariate Spatial Data Hierarchical Modelling for Univariate and Multivariate Spatial Data p. 1/4 Hierarchical Modelling for Univariate and Multivariate Spatial Data Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota

More information

Computer Model Calibration: Bayesian Methods for Combining Simulations and Experiments for Inference and Prediction

Computer Model Calibration: Bayesian Methods for Combining Simulations and Experiments for Inference and Prediction Computer Model Calibration: Bayesian Methods for Combining Simulations and Experiments for Inference and Prediction Example The Lyon-Fedder-Mobary (LFM) model simulates the interaction of solar wind plasma

More information

Gaussian Processes and Complex Computer Models

Gaussian Processes and Complex Computer Models Gaussian Processes and Complex Computer Models Astroinformatics Summer School, Penn State University June 2018 Murali Haran Department of Statistics, Penn State University Murali Haran, Penn State 1 Modeling

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department

More information

Bayesian Learning in Undirected Graphical Models

Bayesian Learning in Undirected Graphical Models Bayesian Learning in Undirected Graphical Models Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK http://www.gatsby.ucl.ac.uk/ Work with: Iain Murray and Hyun-Chul

More information

Lecture 4: Types of errors. Bayesian regression models. Logistic regression

Lecture 4: Types of errors. Bayesian regression models. Logistic regression Lecture 4: Types of errors. Bayesian regression models. Logistic regression A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting more generally COMP-652 and ECSE-68, Lecture

More information

Introduction to Gaussian Processes

Introduction to Gaussian Processes Introduction to Gaussian Processes Neil D. Lawrence GPSS 10th June 2013 Book Rasmussen and Williams (2006) Outline The Gaussian Density Covariance from Basis Functions Basis Function Representations Constructing

More information

Gaussian Processes for Computer Experiments

Gaussian Processes for Computer Experiments Gaussian Processes for Computer Experiments Jeremy Oakley School of Mathematics and Statistics, University of Sheffield www.jeremy-oakley.staff.shef.ac.uk 1 / 43 Computer models Computer model represented

More information

Regression. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh.

Regression. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. Regression Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh September 24 (All of the slides in this course have been adapted from previous versions

More information

The Bayesian approach to inverse problems

The Bayesian approach to inverse problems The Bayesian approach to inverse problems Youssef Marzouk Department of Aeronautics and Astronautics Center for Computational Engineering Massachusetts Institute of Technology ymarz@mit.edu, http://uqgroup.mit.edu

More information

Bayesian Calibration of Inexact Computer Models

Bayesian Calibration of Inexact Computer Models Bayesian Calibration of Inexact Computer Models James Matuk Research Group in Design of Physical and Computer Experiments March 5, 2018 James Matuk (STAT 8750.02) Calibration of Inexact Computer Models

More information

Bayesian model selection for computer model validation via mixture model estimation

Bayesian model selection for computer model validation via mixture model estimation Bayesian model selection for computer model validation via mixture model estimation Kaniav Kamary ATER, CNAM Joint work with É. Parent, P. Barbillon, M. Keller and N. Bousquet Outline Computer model validation

More information

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference 1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Bayesian sensitivity analysis of a cardiac cell model using a Gaussian process emulator Supporting information

Bayesian sensitivity analysis of a cardiac cell model using a Gaussian process emulator Supporting information Bayesian sensitivity analysis of a cardiac cell model using a Gaussian process emulator Supporting information E T Y Chang 1,2, M Strong 3 R H Clayton 1,2, 1 Insigneo Institute for in-silico Medicine,

More information

Better Simulation Metamodeling: The Why, What and How of Stochastic Kriging

Better Simulation Metamodeling: The Why, What and How of Stochastic Kriging Better Simulation Metamodeling: The Why, What and How of Stochastic Kriging Jeremy Staum Collaborators: Bruce Ankenman, Barry Nelson Evren Baysal, Ming Liu, Wei Xie supported by the NSF under Grant No.

More information

Introduction to emulators - the what, the when, the why

Introduction to emulators - the what, the when, the why School of Earth and Environment INSTITUTE FOR CLIMATE & ATMOSPHERIC SCIENCE Introduction to emulators - the what, the when, the why Dr Lindsay Lee 1 What is a simulator? A simulator is a computer code

More information

STA414/2104 Statistical Methods for Machine Learning II

STA414/2104 Statistical Methods for Machine Learning II STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements

More information

Nonparameteric Regression:

Nonparameteric Regression: Nonparameteric Regression: Nadaraya-Watson Kernel Regression & Gaussian Process Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro,

More information

Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13)

Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) 1. Weighted Least Squares (textbook 11.1) Recall regression model Y = β 0 + β 1 X 1 +... + β p 1 X p 1 + ε in matrix form: (Ch. 5,

More information

Nearest Neighbor Gaussian Processes for Large Spatial Data

Nearest Neighbor Gaussian Processes for Large Spatial Data Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns

More information

Challenges In Uncertainty, Calibration, Validation and Predictability of Engineering Analysis Models

Challenges In Uncertainty, Calibration, Validation and Predictability of Engineering Analysis Models Challenges In Uncertainty, Calibration, Validation and Predictability of Engineering Analysis Models Dr. Liping Wang GE Global Research Manager, Probabilistics Lab Niskayuna, NY 2011 UQ Workshop University

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

Bayesian Calibration of Simulators with Structured Discretization Uncertainty

Bayesian Calibration of Simulators with Structured Discretization Uncertainty Bayesian Calibration of Simulators with Structured Discretization Uncertainty Oksana A. Chkrebtii Department of Statistics, The Ohio State University Joint work with Matthew T. Pratola (Statistics, The

More information

Dynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano

Dynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Lawrence J Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Problem: the time series dimension of data is relatively

More information

Bayesian Gaussian Process Regression

Bayesian Gaussian Process Regression Bayesian Gaussian Process Regression STAT8810, Fall 2017 M.T. Pratola October 7, 2017 Today Bayesian Gaussian Process Regression Bayesian GP Regression Recall we had observations from our expensive simulator,

More information

ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS

ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, N.C.,

More information

ST 740: Markov Chain Monte Carlo

ST 740: Markov Chain Monte Carlo ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:

More information

Hierarchical Modelling for Multivariate Spatial Data

Hierarchical Modelling for Multivariate Spatial Data Hierarchical Modelling for Multivariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Point-referenced spatial data often come as

More information

Approximate Principal Components Analysis of Large Data Sets

Approximate Principal Components Analysis of Large Data Sets Approximate Principal Components Analysis of Large Data Sets Daniel J. McDonald Department of Statistics Indiana University mypage.iu.edu/ dajmcdon April 27, 2016 Approximation-Regularization for Analysis

More information

Hierarchical Modeling for Multivariate Spatial Data

Hierarchical Modeling for Multivariate Spatial Data Hierarchical Modeling for Multivariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department

More information

Eco517 Fall 2014 C. Sims MIDTERM EXAM

Eco517 Fall 2014 C. Sims MIDTERM EXAM Eco57 Fall 204 C. Sims MIDTERM EXAM You have 90 minutes for this exam and there are a total of 90 points. The points for each question are listed at the beginning of the question. Answer all questions.

More information

Geostatistical Modeling for Large Data Sets: Low-rank methods

Geostatistical Modeling for Large Data Sets: Low-rank methods Geostatistical Modeling for Large Data Sets: Low-rank methods Whitney Huang, Kelly-Ann Dixon Hamil, and Zizhuang Wu Department of Statistics Purdue University February 22, 2016 Outline Motivation Low-rank

More information

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes COMP 55 Applied Machine Learning Lecture 2: Gaussian processes Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~hvanho2/comp55

More information

Bayesian tsunami fragility modeling considering input data uncertainty

Bayesian tsunami fragility modeling considering input data uncertainty Bayesian tsunami fragility modeling considering input data uncertainty Raffaele De Risi Research Associate University of Bristol United Kingdom De Risi, R., Goda, K., Mori, N., & Yasuda, T. (2016). Bayesian

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Spatial omain Hierarchical Modelling for Univariate Spatial ata Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A.

More information

Fitting Narrow Emission Lines in X-ray Spectra

Fitting Narrow Emission Lines in X-ray Spectra Outline Fitting Narrow Emission Lines in X-ray Spectra Taeyoung Park Department of Statistics, University of Pittsburgh October 11, 2007 Outline of Presentation Outline This talk has three components:

More information

Introduction to Probabilistic Machine Learning

Introduction to Probabilistic Machine Learning Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning

More information

Bayes Model Selection with Path Sampling: Factor Models

Bayes Model Selection with Path Sampling: Factor Models with Path Sampling: Factor Models Ritabrata Dutta and Jayanta K Ghosh Purdue University 07/02/11 Factor Models in Applications Factor Models in Applications Factor Models Factor Models and Factor analysis

More information

Introduction to Bayesian Inference

Introduction to Bayesian Inference University of Pennsylvania EABCN Training School May 10, 2016 Bayesian Inference Ingredients of Bayesian Analysis: Likelihood function p(y φ) Prior density p(φ) Marginal data density p(y ) = p(y φ)p(φ)dφ

More information

When using physical experimental data to adjust, or calibrate, computer simulation models, two general

When using physical experimental data to adjust, or calibrate, computer simulation models, two general A Preposterior Analysis to Predict Identifiability in Experimental Calibration of Computer Models Paul D. Arendt Northwestern University, Department of Mechanical Engineering 2145 Sheridan Road Room B214

More information

Uncertainty Quantification for Inverse Problems. November 7, 2011

Uncertainty Quantification for Inverse Problems. November 7, 2011 Uncertainty Quantification for Inverse Problems November 7, 2011 Outline UQ and inverse problems Review: least-squares Review: Gaussian Bayesian linear model Parametric reductions for IP Bias, variance

More information

A short introduction to INLA and R-INLA

A short introduction to INLA and R-INLA A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk

More information

Calibrating Environmental Engineering Models and Uncertainty Analysis

Calibrating Environmental Engineering Models and Uncertainty Analysis Models and Cornell University Oct 14, 2008 Project Team Christine Shoemaker, co-pi, Professor of Civil and works in applied optimization, co-pi Nikolai Blizniouk, PhD student in Operations Research now

More information

Machine Learning Srihari. Probability Theory. Sargur N. Srihari

Machine Learning Srihari. Probability Theory. Sargur N. Srihari Probability Theory Sargur N. Srihari srihari@cedar.buffalo.edu 1 Probability Theory with Several Variables Key concept is dealing with uncertainty Due to noise and finite data sets Framework for quantification

More information

Learning about physical parameters: The importance of model discrepancy

Learning about physical parameters: The importance of model discrepancy Learning about physical parameters: The importance of model discrepancy Jenný Brynjarsdóttir 1 and Anthony O Hagan 2 June 9, 2014 Abstract Science-based simulation models are widely used to predict the

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Gaussian Processes (10/16/13)

Gaussian Processes (10/16/13) STA561: Probabilistic machine learning Gaussian Processes (10/16/13) Lecturer: Barbara Engelhardt Scribes: Changwei Hu, Di Jin, Mengdi Wang 1 Introduction In supervised learning, we observe some inputs

More information

Regression with correlation for the Sales Data

Regression with correlation for the Sales Data Regression with correlation for the Sales Data Scatter with Loess Curve Time Series Plot Sales 30 35 40 45 Sales 30 35 40 45 0 10 20 30 40 50 Week 0 10 20 30 40 50 Week Sales Data What is our goal with

More information

Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis

Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis Stéphanie Allassonnière CIS, JHU July, 15th 28 Context : Computational Anatomy Context and motivations :

More information

New Insights into History Matching via Sequential Monte Carlo

New Insights into History Matching via Sequential Monte Carlo New Insights into History Matching via Sequential Monte Carlo Associate Professor Chris Drovandi School of Mathematical Sciences ARC Centre of Excellence for Mathematical and Statistical Frontiers (ACEMS)

More information

Outline Lecture 2 2(32)

Outline Lecture 2 2(32) Outline Lecture (3), Lecture Linear Regression and Classification it is our firm belief that an understanding of linear models is essential for understanding nonlinear ones Thomas Schön Division of Automatic

More information

Dimension Reduction in Abundant High Dimensional Regressions

Dimension Reduction in Abundant High Dimensional Regressions Dimension Reduction in Abundant High Dimensional Regressions Dennis Cook University of Minnesota 8th Purdue Symposium June 2012 In collaboration with Liliana Forzani & Adam Rothman, Annals of Statistics,

More information

Computer Practical: Metropolis-Hastings-based MCMC

Computer Practical: Metropolis-Hastings-based MCMC Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov

More information

Approximate Message Passing

Approximate Message Passing Approximate Message Passing Mohammad Emtiyaz Khan CS, UBC February 8, 2012 Abstract In this note, I summarize Sections 5.1 and 5.2 of Arian Maleki s PhD thesis. 1 Notation We denote scalars by small letters

More information

Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment

Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment Ben Shaby SAMSI August 3, 2010 Ben Shaby (SAMSI) OFS adjustment August 3, 2010 1 / 29 Outline 1 Introduction 2 Spatial

More information

Sequential Monte Carlo Methods for Bayesian Computation

Sequential Monte Carlo Methods for Bayesian Computation Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter

More information

Uncertainty quantification for spatial field data using expensive computer models: refocussed Bayesian calibration with optimal projection

Uncertainty quantification for spatial field data using expensive computer models: refocussed Bayesian calibration with optimal projection University of Exeter Department of Mathematics Uncertainty quantification for spatial field data using expensive computer models: refocussed Bayesian calibration with optimal projection James Martin Salter

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

The Metropolis-Hastings Algorithm. June 8, 2012

The Metropolis-Hastings Algorithm. June 8, 2012 The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings

More information

Vast Volatility Matrix Estimation for High Frequency Data

Vast Volatility Matrix Estimation for High Frequency Data Vast Volatility Matrix Estimation for High Frequency Data Yazhen Wang National Science Foundation Yale Workshop, May 14-17, 2009 Disclaimer: My opinion, not the views of NSF Y. Wang (at NSF) 1 / 36 Outline

More information

Lecture 5: GPs and Streaming regression

Lecture 5: GPs and Streaming regression Lecture 5: GPs and Streaming regression Gaussian Processes Information gain Confidence intervals COMP-652 and ECSE-608, Lecture 5 - September 19, 2017 1 Recall: Non-parametric regression Input space X

More information

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Alan Gelfand 1 and Andrew O. Finley 2 1 Department of Statistical Science, Duke University, Durham, North

More information

ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS

ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS 1. THE CLASS OF MODELS y t {y s, s < t} p(y t θ t, {y s, s < t}) θ t = θ(s t ) P[S t = i S t 1 = j] = h ij. 2. WHAT S HANDY ABOUT IT Evaluating the

More information

Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications

Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications Dongbin Xiu Department of Mathematics, Purdue University Support: AFOSR FA955-8-1-353 (Computational Math) SF CAREER DMS-64535

More information

FREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE

FREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE FREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE Donald A. Pierce Oregon State Univ (Emeritus), RERF Hiroshima (Retired), Oregon Health Sciences Univ (Adjunct) Ruggero Bellio Univ of Udine For Perugia

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

Bayesian techniques for fatigue life prediction and for inference in linear time dependent PDEs

Bayesian techniques for fatigue life prediction and for inference in linear time dependent PDEs Bayesian techniques for fatigue life prediction and for inference in linear time dependent PDEs Marco Scavino SRI Center for Uncertainty Quantification Computer, Electrical and Mathematical Sciences &

More information

Riemann Manifold Methods in Bayesian Statistics

Riemann Manifold Methods in Bayesian Statistics Ricardo Ehlers ehlers@icmc.usp.br Applied Maths and Stats University of São Paulo, Brazil Working Group in Statistical Learning University College Dublin September 2015 Bayesian inference is based on Bayes

More information

Midterm. Introduction to Machine Learning. CS 189 Spring You have 1 hour 20 minutes for the exam.

Midterm. Introduction to Machine Learning. CS 189 Spring You have 1 hour 20 minutes for the exam. CS 189 Spring 2013 Introduction to Machine Learning Midterm You have 1 hour 20 minutes for the exam. The exam is closed book, closed notes except your one-page crib sheet. Please use non-programmable calculators

More information

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota,

More information

Short Questions (Do two out of three) 15 points each

Short Questions (Do two out of three) 15 points each Econometrics Short Questions Do two out of three) 5 points each ) Let y = Xβ + u and Z be a set of instruments for X When we estimate β with OLS we project y onto the space spanned by X along a path orthogonal

More information

Fixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility

Fixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility American Economic Review: Papers & Proceedings 2016, 106(5): 400 404 http://dx.doi.org/10.1257/aer.p20161082 Fixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility By Gary Chamberlain*

More information

Steps in Uncertainty Quantification

Steps in Uncertainty Quantification Steps in Uncertainty Quantification Challenge: How do we do uncertainty quantification for computationally expensive models? Example: We have a computational budget of 5 model evaluations. Bayesian inference

More information

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University FEATURE EXPANSIONS FEATURE EXPANSIONS

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

Big Data Analytics: Optimization and Randomization

Big Data Analytics: Optimization and Randomization Big Data Analytics: Optimization and Randomization Tianbao Yang Tutorial@ACML 2015 Hong Kong Department of Computer Science, The University of Iowa, IA, USA Nov. 20, 2015 Yang Tutorial for ACML 15 Nov.

More information

Fusing point and areal level space-time data. data with application to wet deposition

Fusing point and areal level space-time data. data with application to wet deposition Fusing point and areal level space-time data with application to wet deposition Alan Gelfand Duke University Joint work with Sujit Sahu and David Holland Chemical Deposition Combustion of fossil fuel produces

More information

Least Squares Regression

Least Squares Regression CIS 50: Machine Learning Spring 08: Lecture 4 Least Squares Regression Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may not cover all the

More information

Quantile Regression for Extraordinarily Large Data

Quantile Regression for Extraordinarily Large Data Quantile Regression for Extraordinarily Large Data Shih-Kang Chao Department of Statistics Purdue University November, 2016 A joint work with Stanislav Volgushev and Guang Cheng Quantile regression Two-step

More information

Announcements. Proposals graded

Announcements. Proposals graded Announcements Proposals graded Kevin Jamieson 2018 1 Bayesian Methods Machine Learning CSE546 Kevin Jamieson University of Washington November 1, 2018 2018 Kevin Jamieson 2 MLE Recap - coin flips Data:

More information

Bayesian Learning in Undirected Graphical Models

Bayesian Learning in Undirected Graphical Models Bayesian Learning in Undirected Graphical Models Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK http://www.gatsby.ucl.ac.uk/ and Center for Automated Learning and

More information