Computer Model Calibration: Bayesian Methods for Combining Simulations and Experiments for Inference and Prediction

Similar documents
An introduction to Bayesian statistics and model calibration and a host of related topics

Quantifying Uncertainties in Radiation Hydrodynamics Models

Calibration and validation of computer models for radiative shock experiment

Validating Expensive Simulations with Expensive Experiments: A Bayesian Approach

arxiv: v1 [stat.ap] 13 Aug 2012

Uncertainty quantification and calibration of computer models. February 5th, 2014

Efficiency and Reliability of Bayesian Calibration of Energy Supply System Models

Uncertainty quantification for spatial field data using expensive computer models: refocussed Bayesian calibration with optimal projection

Influence of an intensive UV preionization on evolution and EUV-emission of the laser plasma with Xe gas target (S12)

Bayesian reconstruction of free energy profiles from umbrella samples

Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment

PARAMETER TUNING FOR A MULTI-FIDELITY DYNAMICAL MODEL OF THE MAGNETOSPHERE 1

Part 6: Multivariate Normal and Linear Models

Follow-Up Experimental Designs for Computer Models and Physical Processes

Bayesian linear regression

Challenges In Uncertainty, Calibration, Validation and Predictability of Engineering Analysis Models

Why Study Magnetic Reconnection?

A hierarchical Krylov-Bayes iterative inverse solver for MEG with physiological preconditioning

Probability and Estimation. Alan Moses

Bayesian Models in Machine Learning

Exploring Astrophysical Magnetohydrodynamics Using High-power Laser Facilities

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Climate Change: the Uncertainty of Certainty

Default Priors and Effcient Posterior Computation in Bayesian

Prior information, but no MCMC:

Gaussian Processes for Computer Experiments

Naïve Bayes classification

PREDICTING SOLAR GENERATION FROM WEATHER FORECASTS. Chenlin Wu Yuhan Lou

Annals of Nuclear Energy

Bayesian Analysis of RR Lyrae Distances and Kinematics

Bayesian Inference and MCMC

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours

Bayesian Dynamic Linear Modelling for. Complex Computer Models

Integrated Bayesian Experimental Design

Bayesian Regression Linear and Logistic Regression

GAUSSIAN PROCESS REGRESSION

Hierarchical Bayesian Modeling of Planet Populations. Angie Wolfgang NSF Postdoctoral Fellow, Penn State

Principles of Bayesian Inference

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University

Introduction to Gaussian Processes

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

Principles of Bayesian Inference

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio

Visualization of Xe and Sn Atoms Generated from Laser-Produced Plasma for EUV Light Source

Gaussian with mean ( µ ) and standard deviation ( σ)

Statistical Tools and Techniques for Solar Astronomers

Principles of Bayesian Inference

Bayesian Methods for Machine Learning

Bayesian Calibration of Simulators with Structured Discretization Uncertainty

Bayesian Regression (1/31/13)

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models

Plausible Values for Latent Variables Using Mplus

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability

Non-Parametric Bayes

Primer on statistics:

Bayesian data analysis in practice: Three simple examples

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

STATISTICAL MODELS FOR QUANTIFYING THE SPATIAL DISTRIBUTION OF SEASONALLY DERIVED OZONE STANDARDS

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Point spread function reconstruction from the image of a sharp edge

Learning about physical parameters: The importance of model discrepancy

Lecture 4: Dynamic models

Adaptive Data Assimilation and Multi-Model Fusion

Introduction to Gaussian Processes

DRAGON ADVANCED TRAINING COURSE IN ATMOSPHERE REMOTE SENSING. Inversion basics. Erkki Kyrölä Finnish Meteorological Institute

Parameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets

Using a Relativistic Electron Beam to Generate Warm Dense Matter for Equation of State Studies

Chapter 8: Sampling distributions of estimators Sections

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Sampling and incomplete network data

Parametric Techniques Lecture 3

Advanced Statistical Methods. Lecture 6

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Advanced Machine Learning

Multi-fidelity co-kriging models

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Statistical Techniques in Robotics (16-831, F12) Lecture#21 (Monday November 12) Gaussian Processes

Model comparison. Christopher A. Sims Princeton University October 18, 2016

Tracking properties of the ATLAS Transition Radiation Tracker (TRT)

Statistical Techniques in Robotics (16-831, F12) Lecture#20 (Monday November 12) Gaussian Processes

Classical and Bayesian inference

Parametric Techniques

Bayesian Gaussian Process Regression

Using training sets and SVD to separate global 21-cm signal from foreground and instrument systematics

SURROGATE PREPOSTERIOR ANALYSES FOR PREDICTING AND ENHANCING IDENTIFIABILITY IN MODEL CALIBRATION

Data Assimilation: Finding the Initial Conditions in Large Dynamical Systems. Eric Kostelich Data Mining Seminar, Feb. 6, 2006

David Giles Bayesian Econometrics

Problem set: solar irradiance and solar wind

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Bayesian dynamic modeling for large space-time weather datasets using Gaussian predictive processes

Abel Inversion using the Maximum Entropy Method

Wrapped Gaussian processes: a short review and some new results

Models for models. Douglas Nychka Geophysical Statistics Project National Center for Atmospheric Research

Linear Models A linear model is defined by the expression

Multistage Anslysis on Solar Spectral Analyses with Uncertainties in Atomic Physical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Three-dimensional nature of magnetic reconnection X-line in asymmetric current sheets

Combining Experimental Data and Computer Simulations, With an Application to Flyer Plate Experiments

Transcription:

Computer Model Calibration: Bayesian Methods for Combining Simulations and Experiments for Inference and Prediction

Example The Lyon-Fedder-Mobary (LFM) model simulates the interaction of solar wind plasma in the magnetosphere We see this as the Aurora Borealis Have a computer model that attempts to capture the main features of this phenomenon Outputs are large 3-dimensional space-time fields or extracted features thereof Computer model has three inputs: =(,, R)

Example Field observations available from the Polar Ultraviolet Imager satellite Computer model is assumed to capture all salient features of the solar wind interactions, up to random error However, the inputs ( =(,, R) ) are not known Scientific problem: estimate =(,, R)

Basic inverse problem Statistical formulation y s (t) = (t) y f ( ) = ( )+ Where, y f system response y s simulator response at input t θ calibration parameters ε random error Have data from 2 separate sources computer model and field observations Problem is to estimate the calibration parameters

Solutions Many different solutions to such problems can you think of one? Suppose the computer model is fast? Suppose the computer model is slow?

Model calibration Statistical formulation y s (x, t) = (x, t) y f (x, ) = (x, )+ Where, x model or system inputs; y f system response y s simulator response θ calibration parameters ε random error Have data from 2 separate sources computer model and field observations Problem is to estimate the calibration parameters

Solutions Many different solutions to such problems can you think of one? Suppose the computer model is fast? Suppose the computer model is slow?

Example: Radiative shock experiment A radiative shock is a wave in which both hydrodynamic and radiation transport physics play a significant role in the shock s propagation and structure

Application Initial experiments:" 1 ns, 3.8 kj laser irradiates Be diskè plasma down Xe gas filled shock tube at ~ 200 km/s" Circular tube; diam = 575μm" Timing 13-14 ns" Additional experiments " Laser energy ~ 3.8kJ" Circular tube; diam = 575, 1150μm" Timing 13,20,26 ns" Nozzle experiments" Laser energy ~ 3.8kJ" nozzle length = taper length = 500μm" Circular tube; diam = 575μm" Timing 26 ns" " Extrapolation experiments (5 th year) " Laser energy ~ 3.8kJ" Elliptical tube; diam = 575-1150μm" Aspect ratio = 2" Includes nozzle in shock tube" Timing 26 ns" Laser Laser Laser Laser Be disk shock tube 575μm tube filled with Xe gas ~1.2atm shock tube 1150μm nozzle length taper length 575μm 1150μm shock tube 1150μm 575μm

Have several outputs & inputs y Outputs ( ) Shock location Shock breakout time Wall shock location Axial centroid of Xe Area of dense Xe Shock location x Inputs ( ) Observation time Laser energy Be disk thickness Xe gas pressure Tube geometry Calibration parameters ( ) Vary with model Electron flux limiter Laser scale factor Fixed window Centroid of dense Xe Wall shock location Area of dense Xe

We can measure and we can compute 600 µm 1200 µm Circular Elliptical tube tube nozzle nozzle Goal is to predict elliptical tube quantities of interest and uncertainty, without using any data from elliptical Department tube of Statistics experiments and Actuarial Science Shocks at 13 ns

Data Have observations from 1-D CRASH model and experiments Experiment data: 9 experiments experiment variables: Be thickness, Laser Energy, Xe pressure and Time responses: Shock location 1-D CRASH Simulations 320 simulations, varied over 8 inputs experiment variables: Be thickness, Laser Energy, Xe pressure and Time calibration parameters: Be Gamma, Be OSF, Xe Gamma, Xe OSF response: Shock location

Use observations and simulations for prediction in real world Have simulations from the CRASH code Have observations from laboratory experiments Want to combine these sources of data to make prediction of the real-world process... Also have to estimate the calibration parameters Additional complicating factor, the computer model is not an exact representation of the mean of the physical system darn! Approach: Model calibration (Kennedy and O Hagan, 2001; Higdon et al., 2004)

Model calibration Statistical formulation y s y f ( x,t) = η x,t ( ) ( x,θ) = η( x,θ) +δ( x) +ε Where, x model or system inputs; y f system response simulator response y s θ calibration parameters ε random error

Model calibration Statistical formulation y s y f ( x,t) = η x,t ( ) ( x,θ) = η( x,θ) +δ( x) +ε Shared signal

Model calibration Statistical formulation y s y f ( x,t) = η x,t ( ) ( x,θ) = η( x,θ) +δ( x) +ε Discrepancy

Model calibration Statistical formulation y s y f ( x,t) = η x,t ( ) ( x,θ) = η( x,θ) +δ( x) +ε Gaussian process models

Back to Gaussian process models y(x i )=µ + z(x i ) 2 z(x i ) N(0, z) cor ((z(x i ),z(x j )) = e 2 z(x) N(0 n, z R) 2 y(x) N(µ1 n, z R) P d k=1 k(x ik x jk ) 2

Gaussian process model for the emulator in this setting y s (x i,t i )=µ + z(x i,t i ) z(x i,t i ) N(0, 2 z) cor ((z(x i ),z(x j )) = e z(x, t) N(0 n, 2 z R) y(x, t) N(µ1 n, 2 z R) P d k=1 k(x ik x jk ) 2 P d 0 k=1! k(t ik t jk ) 2

Have Gaussian process model for the discrepancy (x i ) N(0, 2 ) P d k=1 k(x ik x jk ) 2 cor (( (x i ), (x j )) = e (x) N(0 m, 2 R )=N(0 m, )

Hierarchical model is used to combine simulations and observations View computational model as a draw of a random process (like before) Denote vectors of simulation trials as and field measurements as y s & y f Suppose that these are n and m- vectors respectively Can combine sources of information using a single GP y s y f " $ y = $ $ # ( x,t) = η x,t ( ) ( x,θ) = η( x,θ) +δ( x) +ε y s y f % ' ' ' & ~ N ( µ, Σ η + Σ δ + Σ ε )

Hierarchical model is used to combine simulations and observations View computational model as a draw of a random process Denote vectors of simulation trials as and field measurements as y s & y f Suppose that these are n and m- vectors respectively Can combine sources of information using a single GP y s y f " $ y = $ $ # ( x,t) = η x,t ( ) ( x,θ) = η( x,θ) +δ( x) +ε y s y f % ' ' ' & ~ N ( µ, Σ η + Σ δ + Σ ε ) Contains correlations between all sources of data via the joint signal in the computer model

Hierarchical model is used to combine simulations and observations View computational model as a draw of a random process Denote vectors of simulation trials as and field measurements as y s & y f Suppose that these are n and m- vectors respectively Can combine sources of information using a single GP y s y f " $ y = $ $ # ( x,t) = η x,t ( ) ( x,θ) = η( x,θ) +δ( x) +ε y s y f % ' ' ' & ~ N ( µ, Σ η + Σ δ + Σ ε ) Problem correlations involving field trials have t =

Hierarchical model is used to combine simulations and observations View computational model as a draw of a random process Denote vectors of simulation trials as and field measurements as y s & y f Suppose that these are n and m- vectors respectively Can combine sources of information using a single GP y s y f " $ y = $ $ # ( x,t) = η x,t ( ) ( x,θ) = η( x,θ) +δ( x) +ε y s y f % ' ' ' & ~ N ( µ, Σ η + Σ δ + Σ ε ) Only operates on the field data

Calibration idea: No discrepancy model (Plot taken shamelessly from Dave Higdon)

Calibration idea: Discrepancy model (Plot taken shamelessly from Dave Higdon) (a) model runs (b) data & prior uncertainty (c) posterior mean for!(x,t) 3!(x,t) 2 0!2 y(x),!(x,t) 2 1 0!1!(x,t) 3 2 1 0!1 0 0.5 x 1 1 t 0.5 0!2 0.5 1 "!3 0 0.5 1 x 0 0.5 x 1 1 t 0.5 0 (d) calibrated simulator prediction (e) posterior model discrepancy (f) calibrated prediction 3 3 3 2 2 2 y(x),!(x,") 1 0!1 #(x) 1 0!1 y(x), $(x) 1 0!1!2 0.5 1 "!3 0 0.5 1 x!2!3 0 0.5 1 x!2!3 0 0.5 1 x

Comments Interpretation of unknown constants can change Without discrepancy: estimating unknown physical constants With discrepancy: selecting tuning constants that best fit the model

Estimation What parameters do we have to estimate?

Estimation Can apply Bayes Rule: In the unbiased case for a fast simulator, For the simulators we have been considering, use an emulator in place of computer model basic idea [ ]! [ˆ, y s ] [A B] = [B A][A] [B] [, y f ] / [y f, ][ ][ ] Ignoring statistical model parameters for the moment [ ˆ, y f ] / [y f ˆ, ][ˆ, y s ][ ]

Estimation How could we put this together for the CRASH problem?

Back to CRASH Inverted-gamma priors for variance components Gamma priors for the correlation parameters Log-normal priors for the calibration parameters Samples from the posterior are generated through MCMC anyone know how to do this?

CRASH predictions

Calibration - BE Gamma

Calibration - BE OSF

Calibration - XE Gamma

Calibration - XE OSF

Exploring the discrepancy The CRASH project was an ongoing endeavor Computer codes were being developed Idea is to use the statistical model to inform code development Predicted the discrepancy over a 4-d grid Used posterior mean surface

Discrepancy

Reasons for the discrepancy It was interesting that the discrepancy is always positive Incorrect radial loss of energy results in Xe that is too hot in front of the shock, and that systematically messes up the shock speed The region where the discrepancy is highest is thought to be the region where Xe pressure is largely going to impact Why might this be a crazy approach to explore the discrepancy?