Calibration and uncertainty quantification using multivariate simulator output
|
|
- Jodie Sanders
- 5 years ago
- Views:
Transcription
1 Calibration and uncertainty quantification using multivariate simulator output LA-UR 4-74 Dave Higdon, Statistical Sciences Group, LANL Jim Gattiker, Statistical Sciences Group, LANL Brian Williams, Statistical Sciences Group, LANL
2 Inference combining a physics model with experimental data drop time drop height (floor) Data generated from model: d z =.3 dz dt dt + ǫ simulation model: d z = dt drop time statistical model: y(z) = η(z) + δ(z) + ǫ drop height (floor) drop time Improved physics model: d z = θ dz dt dt + ǫ statistical model: y(z) = η(z, θ) + δ(z) + ǫ drop height (floor)
3 Basic formulation borrows from Kennedy and O Hagan () 7 (a) (b) (c) (d) η(x), y(x). η(x), y(x). δ(x). y(x) x 3. x x x experimental conditions θ calibration parameters ζ(x) true physical system response given inputs x η(x, θ) simulator response at x and θ. y(x) experimental observation of the physical system δ(x) discrepancy between ζ(x) and η(x, θ) may be decomposed into numerical error and bias e(x) observation error of the experimental data y(x) = ζ(x) + e(x) y(x) = η(x, θ) + δ(x) + e(x) 3. x
4 A Bayesian approach for combining simulations and experimental data for forecasting, calibration and uncertainty quantification prior uncertainty posterior uncertainty 3 3 y(x), η(x,θ) y(x), η(x,θ). θ x. θ x A simple example... x model or system inputs θ model calibration parameters ζ(x) true physical system response given inputs x η(x, θ) simulator response at x and θ. y(x) experimental observation of the physical system e(x) observation error of the experimental data Assume: y(x) = ζ(x) + e(x) = η(x, θ) + e(x) θ unknown. Standard Bayesian estimation gives: π(θ y(x)) L(y(x) η(x, θ)) π(θ)
5 Accounting for limited simulator runs y(x), η(x,θ) 3 prior uncertainty. θ 3. x Borrows from Kennedy and O Hagan (). x model or system inputs θ calibration parameters ζ(x) true physical system response given inputs x η(x, θ) simulator response at x and θ. simulator run at limited input settings η = (η(x, θ),..., η(x m, θm)) T treat η(, ) as a random function use GP prior for η(, ) y(x) experimental observation of the physical system e(x) observation error of the experimental data y(x) = ζ(x) + e(x) y(x) = η(x, θ) + e(x)
6 Accounting for limited simulation runs prior uncertainty posterior realizations of η(x,t) posterior uncertainty 3 3 y(x), η(x,θ). θ 3. x Again, standard Bayesian estimation gives: η(x,t). x π(θ, η(, ) y(x)) L(y(x) η(x, θ)) π(θ) π(η(, )) t. y(x), η(x,θ). θ 3. x Posterior means and quantiles shown. Uncertainty in θ and η(x, θ) are incorporated into the forecast. Gaussian process models for η(, ).
7 Accounting for model discrepancy prior uncertainty Borrows from Kennedy and O Hagan (). y(x), η(x,θ) 3. θ 3. x x model or system inputs θ model or system inputs ζ(x) true physical system response given inputs x η(x, θ) simulator response at x and θ. y(x) experimental observation of the physical system δ(x) discrepancy between ζ(x) and η(x, θ) may be decomposed into numerical error and bias e(x) observation error of the experimental data y(x) = ζ(x) + e(x) y(x) = η(x, θ) + δ(x) + e(x) y(x) = η(x, θ) + δ n (x) + δ b (x) + e(x)
8 Accounting for model discrepancy prior uncertainty posterior model uncertainty y(x), η(x,θ) δ(x) 3. θ 3. x 3 posterior model discrepancy 3. x y(x), η(x,θ) y(x) 3. θ 3. x 3 calibrated forecast 3. x Again, standard Bayesian estimation gives: π(θ, δ n, δ b y(x)) L(y(x) η(x, θ), δ(x)) π(θ) π(η) π(δ) Posterior means and 9% CI s shown. Posterior prediction for ζ(x) is obtained by computing the posterior distribution for η(x, θ) + δ(x) Uncertainty in θ, η(x, t), and δ(x) are incorporated into the forecast. Gaussian process models for η(x, t) and δ(x)
9 Application: implosions of steel cylinders Neddermeyer 43 Initial work on implosion for fat man. Use high explosive (HE) to crush steel cylindrical shells Investigate the feasability of a controlled implosion
10 Some History Early work on cylinders called beer can experiments. Early work not encouraging:...i question Dr. Neddermeyer s seriousness... Deke Parsons. It stinks. R. Feynman Teller and VonNeumann were quite supportive of the implosion idea Data on collapsing cylinder from high speed photography. Symmetrical implosion eventually accomplished using HE lenses by Kistiakowsky. Implosion played a key role in early computer experiments. Feynman worked on implosion calculations with IBM accounting machines. Eventually first computer with addressable memory was developed (MANIAC ).
11 The Experiments
12 Neddermeyer s Model cm 3 4 X s 3 4 X s 3 4 X s 3 4 X s 3 4 X s cm 3 4 X s 3 4 X s 3 4 X s 3 4 X s 3 4 X s Energy from HE imparts an initial inward velocity to the cylinder v = m e u m + m e /m mass ratio m e /m of HE to steel; u energy per unit mass from HE. Energy converts to work done on the cylinder: s { work per unit mass = w = r ρ( λ) i log ri ro log ro + λ log λ } r i = scaled inner radius; r o = scaled outer radius; λ = initial r i /r o ; s = steel yielding stress; ρ = density of steel.
13 Neddermeyer s Model cm 3 4 X s 3 4 X s 3 4 X s 3 4 X s 3 4 X s where cm 3 4 X s 3 4 X s 3 4 X s 3 4 X s 3 4 X s ODE: dr dt = [ R f(r) { v s ρ g(r) }] r = inner radius of cylinder varies with time R = initial outer ( radius of cylinder f(r) = r r λ ln + λ ) r g(r) = ( λ ) [r lnr (r + λ ) ln(r + λ ) λ lnλ ] λ = initial ratio of cylinder r(t = )/R constant volume condition: r outer r = λ
14 Goal: use experimental data to calibrate s and u ; obtain prediction uncertainty for new experiment expt cm expt 3 t = µs cm expt t = µs cm t = 4 µs m e /m.3 m e /m.7 m e /m.36 Hypothetical data obtained from photos at different times during the 3 experimental implosions. All cylinders had a.in outer and a.in inner radius. (λ = 3 ).
15 Carry out simulated implosions using Neddermeyer s model Sequence of runs carried at m input settings (x, θ, θ) = (m e /m, s, u ) varying x over predefined ranges using an OA(3, 4 3 θ θ )-based LH design.... x m θm θm.. inner radius (cm). inner radius (cm) time (s) x 3 x 4 time (s) pi angle (radians) pi radius by time radius by time and angle φ. Each simulation produces a n η = 6 vector of radii for times 6 angles.
16 Generating OA-based LH designs Example: N = 6, 3 factors each at 4 levels OA(6, 4 3 ) design induced LH design x_ x x_ x_ x x Ensures some higher dimensional filling relative to standard LH designs.
17 Generating (nearly) OA-based LH designs Example: NOA(48, 4 8 ) N = 48, 8 factors each at 4 levels..4.8 x_ x_ x_ columns of NOA design matrix X are not exactly orthogonal allows more factors with good higher dimensional properties x_4 x_ x_6 x_ x_
18 PC representation of simulation output Ξ = [η ; ;η m ] a n η m matrix that holds output of m simulations SVD decomposition: Ξ = UDV T K η is st p η columns of [ m UD] columns of [ mv T ] have variance Cylinder example: PC (9.4% of variation) PC (.9% of variation) PC 3 (.7% of variation) r.. time angle r.. time angle r.. time angle p η = 3 PC s: K η = [k ;k ;k 3 ] each vector k i holds trace of PC i. k i s do not change with φ from symmetry of Neddermeyer s model. Simulated trace η(x i, θ i, θ i ) = K ηw(x i, θ i, θ i )+ǫ i, ǫ i s iid N(, λ η ), for any set of tried simulation inputs (x i, θ i, θ i ).
19 Gaussian process models for PC weights Want to evaluate η(x, θ, θ ) at arbitrary input setting (x, θ, θ ). Also want analysis to account for uncertainty here. Approach: model each PC weight as a Gaussian process: where w i (x, θ, θ ) GP(, λ wi R((x, θ), (x, θ ); ρ wi )) R((x, θ), (x, θ ); ρ wi ) = Restricting to the design settings p x k= ρ 4(x k x k ) wik p θ k= x θ θ... and specifying x m θm θm w i = (w i (x, θ, θ ),..., w i (x m, θ m, θ m)) T gives iid w i N (, λ wi R((x, θ ); ρ wi ) ), i =,..., p η where R((x, θ ); ρ wi ) m m is given by (). ρ 4(θ k θ k ) wi(k+p x ) () *note: additional nugget term w i iid N (, λ wi R((x, θ ); ρ wi ) + λ ǫi I m), i =,...,pη, may be useful.
20 Gaussian process models for PC weights At the m simulation input settings the mp η -vector w has prior disribution w λ w R((x, θ ); ρ w ) w =. N.,... w pη λ wp η R((x, θ ); ρ wpη ) w N(, Σ w ); note Σ w = I pη λ w R((x, θ ); ρ w ) can break down. Emulator likelihood: η = vec([η(x, θ, θ ); ;η(x m, θ m, θ m)]) mnη L(η w, λ η ) λη exp { λ η(η Kw) T (η Kw) }, λ η Γ(a η, b η ) where n η is the number of observations in a simulated trace and Equivalently K = [I m k ; ;I m k pη ]. L(η w, λ η ) λ mp η η exp { λ η(w ŵ) T (K T K)(w ŵ) } m(nη pη) λη exp { λ ηη T (I K(K T K) K T )η } mpη λη exp { λ η(w ŵ) T (K T K)(w ŵ) }, λ η Γ(a η, b η) a η = a η + m(n η p η ), b η = b η + ηt (I K(K T K) K T )η, ŵ = (K T K) K T η.
21 Gaussian process models for PC weights Resulting posterior can then be based on computed PC weights ŵ: ŵ w,λ η N(w, (λ η K T K) ) w λ w, ρ w N(, Σ w ) ŵ λ η, λ w, ρ w N(, (λ η K T K) + Σ w ) Resulting posterior is then: π(λ η, λ w, ρ w ŵ) (λη K T K) + Σ w exp{ ŵt ([λ η K T K] + Σ w ) ŵ} p η λ a η η e b ηλ η λ a w wi e b wλ wi p η i= p x j= i= ( ρ wij ) b ρ p θ ( ρ wi(j+px )) b ρ j= MCMC via Metropolis works fine here. Bounded range of ρ wij s facilitates MCMC.
22 Posterior distribution of ρ w.8 PC [x θ] PC. 3 [x θ] PC3. 3 [x θ] Separate models by PC More opportunity to take advantage of effect sparsity
23 Predicting simulator output at untried (x, θ, θ ) Want η(x, θ, θ ) = Kw(x, θ, θ ) For a given draw (λ η, λ w, ρ w ) a draw of w can be produced: ( ) (( ) [( ) ]) ŵ (λη K w N, T K) + Σ w,w (λ w, ρ w ) Define Then ( ) V V V = = V V [( ) (λη K T K) ] + Σ w,w (λ w, ρ w ) w ŵ N(V V ŵ, V V V V ) Realizations can be generated from sample of MCMC output. Lots of info (data?) makes conditioning on point estimate ( λ η, λ w, ρ w ) a good approximation to the posterior. Posterior mean or median work well for ( λ η, λ w, ρ w )
24 Exploring sensitivity of simulator output to model inputs Simulator predictions varing input, holding others at nominal
25 Basic formulation borrows from Kennedy and O Hagan () Experiment r, η δ x 4 time x 4 time r, η+δ x 4 time φ φ φ δ x = m e /m.3 θ = s? θ = u? (t, φ) simulation output space x experimental conditions θ calibration parameters ζ(x) true physical system response given conditions x η(x, θ) simulator response at x and θ. y(x) experimental observation of the physical system δ(x) discrepancy between ζ(x) and η(x, θ) may be decomposed into numerical error and bias e(x) observation error of the experimental data y(x) = ζ(x) + e(x) y(x) = η(x, θ) + δ(x) + e(x)
26 Kernel basis representation for spatial processes δ(s) Define p δ basis functions d (s),..., d pδ (s). basis Here d j (s) is normal density cetered at spatial location ω j : d j (s) = π exp{ (s ω j) } set δ(s) = p δ j= s d j (s)v j where v N(, λ v I pδ ). Can represent δ = (δ(s ),..., δ(s n )) T as δ = Dv where D ij = d j (s i )
27 v and d(s) determine spatial processes δ(s) d j (s)v j δ(s) basis... z(s) s s Continuous representation: δ(s) = p δ j= d j (s)v j where v N(, λ v I pδ ). Discrete representation: For δ = (δ(s ),..., δ(s n )) T, δ = Dv where D ij = d j (s i )
28 Basis representation of discrepancy time angle φ Represent discrepancy δ(x) using basis functions and weights p δ = 4 basis functions over (t,φ); D = [d ; ;d pδ ]; d k s hold basis. δ(x) = Dv(x) where v(x) GP (, λ v I pδ R(x, x ;ρ v ) ) with R(x, x ;ρ v ) = p x k= ρ 4(x k x k ) vk ()
29 Integrated model formulation Data y(x ),..., y(x n ) collected for n experiments at input conditions x,..., x n. Each y(x i ) is a collection of n yi measurements over points indexed by (t, φ). y(x i ) = η(x i, θ) + δ(x i ) + e i = K i ( w(x i, θ) + ( D i v(x i ) + ) e i ) v(xi ) y(x i ) w(x i, θ), v(x i ), λ y N [D i ;K i ], (λ w(x i, θ) y W i ) Since support of each y(x i ) varies and doesn t match that of sims, the basis vectors in K i must be interpolated from K η ; similary, D i must be computed from the support of y(x i ): x r. r.. r x time pi pi angle pi angle pi time 4 x pi pi angle time x *note: cubic spline interpolation over (time, φ) used here.
30 Define Integrated model formulation n y = n y + + n yn, the total number of experimental data points, y to be the n y -vector from concatination of the y(x i ) s, v = vec([v(x ); ;v(x n )] T ) and u(θ) = vec([w(x, θ, θ ); ;w(x n, θ, θ )] T ) ( ( ) ) v y v, u(θ), λ y N B, (λ u(θ) y W y ), λ y Γ(a y, b y ) (3) where W y = diag(w,..., W n ) and ( ) P T B = diag(d,..., D n, K,..., K n ) D PK T P D and P K are permutation matricies whose rows are given by: P D (j + n(i ); ) = e T (j )p δ +i, i =,..., p δ; j =,..., n P K (j + n(i ); ) = e T (j )p η +i, i =,..., p η; j =,..., n
31 Integrated model formulation (continued) Equivalently (3) can be represented ) ( ) (( ) ) (ˆv v v, λ û u(θ) y N, (λ u(θ) y B T W y B), λ y Γ(a y, b y) with n y = n y + + n yn, the total number of experimental data points ) (ˆv = (B T W û y B) B T W y y a y = a y + [n y n(p δ + p η )] [ ( b y = b y + )) T ( )) (ˆv (ˆv ] y B W û y y B û model dimension reduction simulator data and discrep standard n η m n y basis p η m n (p δ + p η ) Basis approach particularly efficient when n η and n y are large.
32 λ Marginal likelihood The (marginal) likelihood L(ˆv, û, ŵ λ η, λ w, ρ w, λ y, λ v, ρ v, θ) has the form ˆv Σ v û N, Λ y + ŵ Λ Σ η uw where Λ y = λ y B T W y B Λ η = λ η K T K Σ v = λ v I pη R(x, x;ρ v ) R(x, x;ρ v ) = n n correlation matrix from applying () to the conditions x,..., x n corresponding the the n experiments. Σ uw = w R((x, θ), (x, θ); ρ w) λ w R((x, θ), (x, θ ); ρ w ) λ wp η R((x, θ), (x, θ); ρ wpη ) λ wp η R((x, θ), (x, θ ); ρ wpη ) w R((x, θ ), (x, θ); ρ w ) λ w R((x, θ ), (x, θ ); ρ w ) λ wp η R((x, θ ), (x, θ); ρ wpη ) λ wp η R((x, θ ), (x, θ ); ρ wpη ) λ Permutation of Σ uw is block diagonal can speed up computations. Only off diagonal blocks of Σ uw depend on θ.
33 Posterior distribution Likelihood: L(ˆv, û, ŵ λ η, λ w, ρ w, λ y, λ v, ρ v, θ) Prior: π(λ η, λ w, ρ w, λ y, λ v, ρ v, θ) Posterior: π(λ η, λ w, ρ w, λ y, λ v, ρ v, θ ˆv, û, ŵ) L(ˆv, û, ŵ λ η, λ w, ρ w, λ y, λ v, ρ v, θ) π(λ η, λ w, ρ w, λ y, λ v, ρ v, θ) Posterior exploration via MCMC Can take advantage of structure and sparcity to speed up sampling. A useful approximation to speed up posterior evaluation: π(λ η, λ w, ρ w, λ y, λ v, ρ v, θ ˆv, û, ŵ) L(ŵ λ η, λ w, ρ w ) π(λ η, λ w, ρ w ) L(ˆv, û λ η, λ w, ρ w, λ y, λ v, ρ v, θ) π(λ y, λ v, ρ v, θ) In this approximation, experimental data is not used to inform about parameters λ η, λ w, ρ w which govern the simulator process η(x, θ).
34 Posterior distribution of model parameters (θ, θ )
35 Posterior mean decomposition for each experiment Experiment Experiment Experiment 3 r, η r, η r, η x 4 time φ x 4 time φ x 4 time φ δ δ δ x 4 time φ x 4 time φ x 4 time φ r, η+δ r, η+δ r, η+δ x 4 time φ x 4 time φ x 4 time φ
36 Posterior prediction for implosion in each experiment Experiment Experiment Experiment 3 r... x time φ r... x time φ r... x time φ Experiment r... x time φ r
37 9% prediction intervals for implosions at exposure times Experiment Experiment Experiment 3 Time µs Time 7 µs Time µs Time 4 µs Time 4 µs Time 4 µs Predictions from separate analyses which hold data from the experiment being predicted.
38 Quantifying uncertainty for simulation-based forecasts Simulation-based predictions accumulate uncertainty due to: parameter/calibration uncertainty; simulator discrepancy/inadequacy; observation error in data; sparseness in data Limits on dimensionality? Have dealt with up to -dimensional θ. Statistics typically uses the wrong model (eg. a regression line) to explain data. So the framework is nothing new, in principle. Calibrating a model with substantial inadequacy? only slight inadequacy? The slowness of the simulator and high dimensionality make things difficult. Extrapolation is often a goal in such investigations. Generally, the the closer to reality the simulator is, the more it can be trusted for extrapolation. Can this be more rigorously formalized?
39 Application: Ta Flyerplate Experiment PTW model governs features on the visar velocity profile. Use pricipal components (EOF s) to deal with high dimensionality.
40 Simulations and Data.3 velocity km/s..... standardized velocity.. 3 time µs.. 3 time µs PTW calibration parameters with input domains Parameter Description Domain Min Max θ Initial strain hardening rate Material constant in thermal activation κ energy term relates to the temperature dependence Material constant in thermal activation log(γ) energy term relates to the strain rate dependence y Maximum yield stress (at K) y Minimum yield stress ( melting)..8 s Maximum saturation stress (at K).4.36 s Minimum saturation stress ( melting).8.43
41 PC representation of simulation output.. standardized velocity time (µ s) fit using p η = 7% of variation velocity (km/s) time (µ s) standardized velocity.. 3 time (µ s)
42 Marginal Posteriors for spatial correlation parameters ρ wij PC.9.9 PC PC θ θ θ
43 PC - based sensitivities
44 .8 Discrepancy basis stadardized velocity time (µ s) Represent discrepancy δ(x) using basis functions and weights Here d j (s) is normal density cetered at spatial location ω j : d j (s) = exp{ π (s ω j) } set δ(s) = p δ j= d j (s)v j where v N(, λ v I pδ ). Can represent δ = (δ(s ),..., δ(s n )) T as δ = Dv where p δ = basis functions over t. D ij = d j (s i )
45 posterior distribution for PTW parameters
46 posterior predictive distribution for trace η Y time (µs) time (µs) standardized velocity standardized velocity 4 δ time (µs) time (µs) standardized velocity standardized velocity
47 posterior predictive distribution for trace (original scale) η Y velocity. velocity time (µs)... 3 time (µs).3 δ.3.. velocity. velocity time (µs)... 3 time (µs)
Computer Model Calibration using High Dimensional Output
Computer Model Calibration using High Dimensional Output Dave Higdon, Los Alamos National Laboratory Jim Gattiker, Los Alamos National Laboratory Brian Williams, Los Alamos National Laboratory Maria Rightley,
More informationCombining Experimental Data and Computer Simulations, With an Application to Flyer Plate Experiments
Combining Experimental Data and Computer Simulations, With an Application to Flyer Plate Experiments LA-UR-6- Brian Williams, Los Alamos National Laboratory Dave Higdon, Los Alamos National Laboratory
More informationSequential Importance Sampling for Rare Event Estimation with Computer Experiments
Sequential Importance Sampling for Rare Event Estimation with Computer Experiments Brian Williams and Rick Picard LA-UR-12-22467 Statistical Sciences Group, Los Alamos National Laboratory Abstract Importance
More informationUncertainty quantification and calibration of computer models. February 5th, 2014
Uncertainty quantification and calibration of computer models February 5th, 2014 Physical model Physical model is defined by a set of differential equations. Now, they are usually described by computer
More informationBayesian inference & process convolution models Dave Higdon, Statistical Sciences Group, LANL
1 Bayesian inference & process convolution models Dave Higdon, Statistical Sciences Group, LANL 2 MOVING AVERAGE SPATIAL MODELS Kernel basis representation for spatial processes z(s) Define m basis functions
More informationAn introduction to Bayesian statistics and model calibration and a host of related topics
An introduction to Bayesian statistics and model calibration and a host of related topics Derek Bingham Statistics and Actuarial Science Simon Fraser University Cast of thousands have participated in the
More informationGaussian Process Modeling in Cosmology: The Coyote Universe. Katrin Heitmann, ISR-1, LANL
1998 2015 Gaussian Process Modeling in Cosmology: The Coyote Universe Katrin Heitmann, ISR-1, LANL In collaboration with: Jim Ahrens, Salman Habib, David Higdon, Chung Hsu, Earl Lawrence, Charlie Nakhleh,
More informationFast Dimension-Reduced Climate Model Calibration and the Effect of Data Aggregation
Fast Dimension-Reduced Climate Model Calibration and the Effect of Data Aggregation Won Chang Post Doctoral Scholar, Department of Statistics, University of Chicago Oct 15, 2014 Thesis Advisors: Murali
More informationNonlinear Models. and. Hierarchical Nonlinear Models
Nonlinear Models and Hierarchical Nonlinear Models Start Simple Progressively Add Complexity Tree Allometries Diameter vs Height with a hierarchical species effect Three response variables: Ht, crown depth,
More informationEcon 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines
Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the
More informationBayesian Dynamic Linear Modelling for. Complex Computer Models
Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer
More informationSURROGATE PREPOSTERIOR ANALYSES FOR PREDICTING AND ENHANCING IDENTIFIABILITY IN MODEL CALIBRATION
International Journal for Uncertainty Quantification, ():xxx xxx, 0 SURROGATE PREPOSTERIOR ANALYSES FOR PREDICTING AND ENHANCING IDENTIFIABILITY IN MODEL CALIBRATION Zhen Jiang, Daniel W. Apley, & Wei
More information1. Gaussian process emulator for principal components
Supplement of Geosci. Model Dev., 7, 1933 1943, 2014 http://www.geosci-model-dev.net/7/1933/2014/ doi:10.5194/gmd-7-1933-2014-supplement Author(s) 2014. CC Attribution 3.0 License. Supplement of Probabilistic
More informationGaussian processes for uncertainty quantification in computer experiments
Gaussian processes for uncertainty quantification in computer experiments Richard Wilkinson University of Nottingham Gaussian process summer school, Sheffield 2013 Richard Wilkinson (Nottingham) GPs for
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More informationConstruction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting
Construction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting Anne Philippe Laboratoire de Mathématiques Jean Leray Université de Nantes Workshop EDF-INRIA,
More informationHierarchical Modeling for Univariate Spatial Data
Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This
More informationGaussian with mean ( µ ) and standard deviation ( σ)
Slide from Pieter Abbeel Gaussian with mean ( µ ) and standard deviation ( σ) 10/6/16 CSE-571: Robotics X ~ N( µ, σ ) Y ~ N( aµ + b, a σ ) Y = ax + b + + + + 1 1 1 1 1 1 1 1 1 1, ~ ) ( ) ( ), ( ~ ), (
More informationHierarchical Modelling for Univariate and Multivariate Spatial Data
Hierarchical Modelling for Univariate and Multivariate Spatial Data p. 1/4 Hierarchical Modelling for Univariate and Multivariate Spatial Data Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota
More informationComputer Model Calibration: Bayesian Methods for Combining Simulations and Experiments for Inference and Prediction
Computer Model Calibration: Bayesian Methods for Combining Simulations and Experiments for Inference and Prediction Example The Lyon-Fedder-Mobary (LFM) model simulates the interaction of solar wind plasma
More informationGaussian Processes and Complex Computer Models
Gaussian Processes and Complex Computer Models Astroinformatics Summer School, Penn State University June 2018 Murali Haran Department of Statistics, Penn State University Murali Haran, Penn State 1 Modeling
More informationHierarchical Modelling for Univariate Spatial Data
Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department
More informationBayesian Learning in Undirected Graphical Models
Bayesian Learning in Undirected Graphical Models Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK http://www.gatsby.ucl.ac.uk/ Work with: Iain Murray and Hyun-Chul
More informationLecture 4: Types of errors. Bayesian regression models. Logistic regression
Lecture 4: Types of errors. Bayesian regression models. Logistic regression A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting more generally COMP-652 and ECSE-68, Lecture
More informationIntroduction to Gaussian Processes
Introduction to Gaussian Processes Neil D. Lawrence GPSS 10th June 2013 Book Rasmussen and Williams (2006) Outline The Gaussian Density Covariance from Basis Functions Basis Function Representations Constructing
More informationGaussian Processes for Computer Experiments
Gaussian Processes for Computer Experiments Jeremy Oakley School of Mathematics and Statistics, University of Sheffield www.jeremy-oakley.staff.shef.ac.uk 1 / 43 Computer models Computer model represented
More informationRegression. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh.
Regression Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh September 24 (All of the slides in this course have been adapted from previous versions
More informationThe Bayesian approach to inverse problems
The Bayesian approach to inverse problems Youssef Marzouk Department of Aeronautics and Astronautics Center for Computational Engineering Massachusetts Institute of Technology ymarz@mit.edu, http://uqgroup.mit.edu
More informationBayesian Calibration of Inexact Computer Models
Bayesian Calibration of Inexact Computer Models James Matuk Research Group in Design of Physical and Computer Experiments March 5, 2018 James Matuk (STAT 8750.02) Calibration of Inexact Computer Models
More informationBayesian model selection for computer model validation via mixture model estimation
Bayesian model selection for computer model validation via mixture model estimation Kaniav Kamary ATER, CNAM Joint work with É. Parent, P. Barbillon, M. Keller and N. Bousquet Outline Computer model validation
More informationBayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference
1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationBayesian sensitivity analysis of a cardiac cell model using a Gaussian process emulator Supporting information
Bayesian sensitivity analysis of a cardiac cell model using a Gaussian process emulator Supporting information E T Y Chang 1,2, M Strong 3 R H Clayton 1,2, 1 Insigneo Institute for in-silico Medicine,
More informationBetter Simulation Metamodeling: The Why, What and How of Stochastic Kriging
Better Simulation Metamodeling: The Why, What and How of Stochastic Kriging Jeremy Staum Collaborators: Bruce Ankenman, Barry Nelson Evren Baysal, Ming Liu, Wei Xie supported by the NSF under Grant No.
More informationIntroduction to emulators - the what, the when, the why
School of Earth and Environment INSTITUTE FOR CLIMATE & ATMOSPHERIC SCIENCE Introduction to emulators - the what, the when, the why Dr Lindsay Lee 1 What is a simulator? A simulator is a computer code
More informationSTA414/2104 Statistical Methods for Machine Learning II
STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements
More informationNonparameteric Regression:
Nonparameteric Regression: Nadaraya-Watson Kernel Regression & Gaussian Process Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro,
More informationStat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13)
Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) 1. Weighted Least Squares (textbook 11.1) Recall regression model Y = β 0 + β 1 X 1 +... + β p 1 X p 1 + ε in matrix form: (Ch. 5,
More informationNearest Neighbor Gaussian Processes for Large Spatial Data
Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns
More informationChallenges In Uncertainty, Calibration, Validation and Predictability of Engineering Analysis Models
Challenges In Uncertainty, Calibration, Validation and Predictability of Engineering Analysis Models Dr. Liping Wang GE Global Research Manager, Probabilistics Lab Niskayuna, NY 2011 UQ Workshop University
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More informationBayesian Calibration of Simulators with Structured Discretization Uncertainty
Bayesian Calibration of Simulators with Structured Discretization Uncertainty Oksana A. Chkrebtii Department of Statistics, The Ohio State University Joint work with Matthew T. Pratola (Statistics, The
More informationDynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano
Dynamic Factor Models and Factor Augmented Vector Autoregressions Lawrence J Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Problem: the time series dimension of data is relatively
More informationBayesian Gaussian Process Regression
Bayesian Gaussian Process Regression STAT8810, Fall 2017 M.T. Pratola October 7, 2017 Today Bayesian Gaussian Process Regression Bayesian GP Regression Recall we had observations from our expensive simulator,
More informationESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS
ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, N.C.,
More informationST 740: Markov Chain Monte Carlo
ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:
More informationHierarchical Modelling for Multivariate Spatial Data
Hierarchical Modelling for Multivariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Point-referenced spatial data often come as
More informationApproximate Principal Components Analysis of Large Data Sets
Approximate Principal Components Analysis of Large Data Sets Daniel J. McDonald Department of Statistics Indiana University mypage.iu.edu/ dajmcdon April 27, 2016 Approximation-Regularization for Analysis
More informationHierarchical Modeling for Multivariate Spatial Data
Hierarchical Modeling for Multivariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department
More informationEco517 Fall 2014 C. Sims MIDTERM EXAM
Eco57 Fall 204 C. Sims MIDTERM EXAM You have 90 minutes for this exam and there are a total of 90 points. The points for each question are listed at the beginning of the question. Answer all questions.
More informationGeostatistical Modeling for Large Data Sets: Low-rank methods
Geostatistical Modeling for Large Data Sets: Low-rank methods Whitney Huang, Kelly-Ann Dixon Hamil, and Zizhuang Wu Department of Statistics Purdue University February 22, 2016 Outline Motivation Low-rank
More informationCOMP 551 Applied Machine Learning Lecture 20: Gaussian processes
COMP 55 Applied Machine Learning Lecture 2: Gaussian processes Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~hvanho2/comp55
More informationBayesian tsunami fragility modeling considering input data uncertainty
Bayesian tsunami fragility modeling considering input data uncertainty Raffaele De Risi Research Associate University of Bristol United Kingdom De Risi, R., Goda, K., Mori, N., & Yasuda, T. (2016). Bayesian
More informationHierarchical Modelling for Univariate Spatial Data
Spatial omain Hierarchical Modelling for Univariate Spatial ata Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A.
More informationFitting Narrow Emission Lines in X-ray Spectra
Outline Fitting Narrow Emission Lines in X-ray Spectra Taeyoung Park Department of Statistics, University of Pittsburgh October 11, 2007 Outline of Presentation Outline This talk has three components:
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationBayes Model Selection with Path Sampling: Factor Models
with Path Sampling: Factor Models Ritabrata Dutta and Jayanta K Ghosh Purdue University 07/02/11 Factor Models in Applications Factor Models in Applications Factor Models Factor Models and Factor analysis
More informationIntroduction to Bayesian Inference
University of Pennsylvania EABCN Training School May 10, 2016 Bayesian Inference Ingredients of Bayesian Analysis: Likelihood function p(y φ) Prior density p(φ) Marginal data density p(y ) = p(y φ)p(φ)dφ
More informationWhen using physical experimental data to adjust, or calibrate, computer simulation models, two general
A Preposterior Analysis to Predict Identifiability in Experimental Calibration of Computer Models Paul D. Arendt Northwestern University, Department of Mechanical Engineering 2145 Sheridan Road Room B214
More informationUncertainty Quantification for Inverse Problems. November 7, 2011
Uncertainty Quantification for Inverse Problems November 7, 2011 Outline UQ and inverse problems Review: least-squares Review: Gaussian Bayesian linear model Parametric reductions for IP Bias, variance
More informationA short introduction to INLA and R-INLA
A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk
More informationCalibrating Environmental Engineering Models and Uncertainty Analysis
Models and Cornell University Oct 14, 2008 Project Team Christine Shoemaker, co-pi, Professor of Civil and works in applied optimization, co-pi Nikolai Blizniouk, PhD student in Operations Research now
More informationMachine Learning Srihari. Probability Theory. Sargur N. Srihari
Probability Theory Sargur N. Srihari srihari@cedar.buffalo.edu 1 Probability Theory with Several Variables Key concept is dealing with uncertainty Due to noise and finite data sets Framework for quantification
More informationLearning about physical parameters: The importance of model discrepancy
Learning about physical parameters: The importance of model discrepancy Jenný Brynjarsdóttir 1 and Anthony O Hagan 2 June 9, 2014 Abstract Science-based simulation models are widely used to predict the
More informationPart 6: Multivariate Normal and Linear Models
Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of
More informationGaussian Processes (10/16/13)
STA561: Probabilistic machine learning Gaussian Processes (10/16/13) Lecturer: Barbara Engelhardt Scribes: Changwei Hu, Di Jin, Mengdi Wang 1 Introduction In supervised learning, we observe some inputs
More informationRegression with correlation for the Sales Data
Regression with correlation for the Sales Data Scatter with Loess Curve Time Series Plot Sales 30 35 40 45 Sales 30 35 40 45 0 10 20 30 40 50 Week 0 10 20 30 40 50 Week Sales Data What is our goal with
More informationGenerative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis
Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis Stéphanie Allassonnière CIS, JHU July, 15th 28 Context : Computational Anatomy Context and motivations :
More informationNew Insights into History Matching via Sequential Monte Carlo
New Insights into History Matching via Sequential Monte Carlo Associate Professor Chris Drovandi School of Mathematical Sciences ARC Centre of Excellence for Mathematical and Statistical Frontiers (ACEMS)
More informationOutline Lecture 2 2(32)
Outline Lecture (3), Lecture Linear Regression and Classification it is our firm belief that an understanding of linear models is essential for understanding nonlinear ones Thomas Schön Division of Automatic
More informationDimension Reduction in Abundant High Dimensional Regressions
Dimension Reduction in Abundant High Dimensional Regressions Dennis Cook University of Minnesota 8th Purdue Symposium June 2012 In collaboration with Liliana Forzani & Adam Rothman, Annals of Statistics,
More informationComputer Practical: Metropolis-Hastings-based MCMC
Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov
More informationApproximate Message Passing
Approximate Message Passing Mohammad Emtiyaz Khan CS, UBC February 8, 2012 Abstract In this note, I summarize Sections 5.1 and 5.2 of Arian Maleki s PhD thesis. 1 Notation We denote scalars by small letters
More informationApproximate Bayesian computation for spatial extremes via open-faced sandwich adjustment
Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment Ben Shaby SAMSI August 3, 2010 Ben Shaby (SAMSI) OFS adjustment August 3, 2010 1 / 29 Outline 1 Introduction 2 Spatial
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationUncertainty quantification for spatial field data using expensive computer models: refocussed Bayesian calibration with optimal projection
University of Exeter Department of Mathematics Uncertainty quantification for spatial field data using expensive computer models: refocussed Bayesian calibration with optimal projection James Martin Salter
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationThe Metropolis-Hastings Algorithm. June 8, 2012
The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings
More informationVast Volatility Matrix Estimation for High Frequency Data
Vast Volatility Matrix Estimation for High Frequency Data Yazhen Wang National Science Foundation Yale Workshop, May 14-17, 2009 Disclaimer: My opinion, not the views of NSF Y. Wang (at NSF) 1 / 36 Outline
More informationLecture 5: GPs and Streaming regression
Lecture 5: GPs and Streaming regression Gaussian Processes Information gain Confidence intervals COMP-652 and ECSE-608, Lecture 5 - September 19, 2017 1 Recall: Non-parametric regression Input space X
More informationBayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes
Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Alan Gelfand 1 and Andrew O. Finley 2 1 Department of Statistical Science, Duke University, Durham, North
More informationECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS
ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS 1. THE CLASS OF MODELS y t {y s, s < t} p(y t θ t, {y s, s < t}) θ t = θ(s t ) P[S t = i S t 1 = j] = h ij. 2. WHAT S HANDY ABOUT IT Evaluating the
More informationStochastic Collocation Methods for Polynomial Chaos: Analysis and Applications
Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications Dongbin Xiu Department of Mathematics, Purdue University Support: AFOSR FA955-8-1-353 (Computational Math) SF CAREER DMS-64535
More informationFREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE
FREQUENTIST BEHAVIOR OF FORMAL BAYESIAN INFERENCE Donald A. Pierce Oregon State Univ (Emeritus), RERF Hiroshima (Retired), Oregon Health Sciences Univ (Adjunct) Ruggero Bellio Univ of Udine For Perugia
More informationHastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model
UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced
More informationBayesian techniques for fatigue life prediction and for inference in linear time dependent PDEs
Bayesian techniques for fatigue life prediction and for inference in linear time dependent PDEs Marco Scavino SRI Center for Uncertainty Quantification Computer, Electrical and Mathematical Sciences &
More informationRiemann Manifold Methods in Bayesian Statistics
Ricardo Ehlers ehlers@icmc.usp.br Applied Maths and Stats University of São Paulo, Brazil Working Group in Statistical Learning University College Dublin September 2015 Bayesian inference is based on Bayes
More informationMidterm. Introduction to Machine Learning. CS 189 Spring You have 1 hour 20 minutes for the exam.
CS 189 Spring 2013 Introduction to Machine Learning Midterm You have 1 hour 20 minutes for the exam. The exam is closed book, closed notes except your one-page crib sheet. Please use non-programmable calculators
More informationBayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes
Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota,
More informationShort Questions (Do two out of three) 15 points each
Econometrics Short Questions Do two out of three) 5 points each ) Let y = Xβ + u and Z be a set of instruments for X When we estimate β with OLS we project y onto the space spanned by X along a path orthogonal
More informationFixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility
American Economic Review: Papers & Proceedings 2016, 106(5): 400 404 http://dx.doi.org/10.1257/aer.p20161082 Fixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility By Gary Chamberlain*
More informationSteps in Uncertainty Quantification
Steps in Uncertainty Quantification Challenge: How do we do uncertainty quantification for computationally expensive models? Example: We have a computational budget of 5 model evaluations. Bayesian inference
More informationCOMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017
COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University FEATURE EXPANSIONS FEATURE EXPANSIONS
More informationLinear Methods for Prediction
Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we
More informationBig Data Analytics: Optimization and Randomization
Big Data Analytics: Optimization and Randomization Tianbao Yang Tutorial@ACML 2015 Hong Kong Department of Computer Science, The University of Iowa, IA, USA Nov. 20, 2015 Yang Tutorial for ACML 15 Nov.
More informationFusing point and areal level space-time data. data with application to wet deposition
Fusing point and areal level space-time data with application to wet deposition Alan Gelfand Duke University Joint work with Sujit Sahu and David Holland Chemical Deposition Combustion of fossil fuel produces
More informationLeast Squares Regression
CIS 50: Machine Learning Spring 08: Lecture 4 Least Squares Regression Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may not cover all the
More informationQuantile Regression for Extraordinarily Large Data
Quantile Regression for Extraordinarily Large Data Shih-Kang Chao Department of Statistics Purdue University November, 2016 A joint work with Stanislav Volgushev and Guang Cheng Quantile regression Two-step
More informationAnnouncements. Proposals graded
Announcements Proposals graded Kevin Jamieson 2018 1 Bayesian Methods Machine Learning CSE546 Kevin Jamieson University of Washington November 1, 2018 2018 Kevin Jamieson 2 MLE Recap - coin flips Data:
More informationBayesian Learning in Undirected Graphical Models
Bayesian Learning in Undirected Graphical Models Zoubin Ghahramani Gatsby Computational Neuroscience Unit University College London, UK http://www.gatsby.ucl.ac.uk/ and Center for Automated Learning and
More information