Gaussian Processes for Computer Experiments

Similar documents
Introduction to emulators - the what, the when, the why

An introduction to Bayesian statistics and model calibration and a host of related topics

Estimating percentiles of uncertain computer code outputs

Uncertainty quantification and calibration of computer models. February 5th, 2014

Sensitivity analysis in linear and nonlinear models: A review. Introduction

Structural Uncertainty in Health Economic Decision Models

Bayesian sensitivity analysis of a cardiac cell model using a Gaussian process emulator Supporting information

Uncertainty in energy system models

A comparison of polynomial chaos and Gaussian process emulation for uncertainty quantification in computer experiments

Gaussian Process Regression and Emulation

Lecture : Probabilistic Machine Learning

Introduction to Gaussian Processes

Multi-fidelity sensitivity analysis

Introduction to Gaussian Processes

Uncertainty analysis for computer models.

Bayesian Methods for Machine Learning

Learning about physical parameters: The importance of model discrepancy

Sequential Importance Sampling for Rare Event Estimation with Computer Experiments

Gaussian processes for uncertainty quantification in computer experiments

Uncertainty quantification for spatial field data using expensive computer models: refocussed Bayesian calibration with optimal projection

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

Monitoring Wafer Geometric Quality using Additive Gaussian Process

Introduction to Bayesian Learning. Machine Learning Fall 2018

Markov Chain Monte Carlo methods

19 : Slice Sampling and HMC

CPSC 540: Machine Learning

Bayesian model selection: methodology, computation and applications

ST 740: Markov Chain Monte Carlo

arxiv: v1 [stat.me] 24 May 2010

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

Gaussian process regression for Sensitivity analysis

Principles of Bayesian Inference

Bayes: All uncertainty is described using probability.

SENSITIVITY ANALYSIS IN NUMERICAL SIMULATION OF MULTIPHASE FLOW FOR CO 2 STORAGE IN SALINE AQUIFERS USING THE PROBABILISTIC COLLOCATION APPROACH

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

New Insights into History Matching via Sequential Monte Carlo

Markov Chain Monte Carlo

Bayesian Quadrature: Model-based Approximate Integration. David Duvenaud University of Cambridge

Dynamic System Identification using HDMR-Bayesian Technique

Default Priors and Effcient Posterior Computation in Bayesian

Lectures. Variance-based sensitivity analysis in the presence of correlated input variables. Thomas Most. Source:

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Bayesian Prediction of Code Output. ASA Albuquerque Chapter Short Course October 2014

Bayesian Linear Regression

Comparing Non-informative Priors for Estimation and Prediction in Spatial Models

Practical Bayesian Optimization of Machine Learning. Learning Algorithms

Bayesian Statistics Applied to Complex Models of Physical Systems

Making rating curves - the Bayesian approach

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

SENSITIVITY ANALYSIS BY THE USE OF A SURROGATE MODEL IN LB-LOCA: LOFT L2-5 WITH CATHARE-2 V2.5 CODE

Bayesian Inference and MCMC

Introduction. Chapter 1

Automatic Differentiation Equipped Variable Elimination for Sensitivity Analysis on Probabilistic Inference Queries

Density Estimation. Seungjin Choi

Particle Filtering for Data-Driven Simulation and Optimization

Nonparametric Bayesian Methods (Gaussian Processes)

MCMC algorithms for fitting Bayesian models

Probabilistic numerics for deep learning

Expectation Propagation for Approximate Bayesian Inference

BAYESIAN ESTIMATION OF LINEAR STATISTICAL MODEL BIAS

Bayesian linear regression

Machine Learning - MT & 5. Basis Expansion, Regularization, Validation

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes

STA 4273H: Sta-s-cal Machine Learning

Reliability Monitoring Using Log Gaussian Process Regression

Spatial Statistics with Image Analysis. Outline. A Statistical Approach. Johan Lindström 1. Lund October 6, 2016

arxiv: v1 [stat.me] 10 Jul 2009

Learning Gaussian Process Models from Uncertain Data

Sensitivity analysis and calibration of a global aerosol model

20: Gaussian Processes

Principles of Bayesian Inference

Efficiency and Reliability of Bayesian Calibration of Energy Supply System Models

Bayesian Dynamic Linear Modelling for. Complex Computer Models

Bayesian Machine Learning

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, 2013

Global Optimisation with Gaussian Processes. Michael A. Osborne Machine Learning Research Group Department o Engineering Science University o Oxford

Statistics: Learning models from data

Stratified Random Sampling for Dependent Inputs

Use of Design Sensitivity Information in Response Surface and Kriging Metamodels

Notes on Machine Learning for and

17 : Markov Chain Monte Carlo

Gaussian Process Approximations of Stochastic Differential Equations

STA414/2104 Statistical Methods for Machine Learning II

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, What about continuous variables?

Recap from previous lecture

Professor David B. Stephenson

Emulation of Numerical Models with Over-specified Basis Functions

Integrated Non-Factorized Variational Inference

ADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING. Non-linear regression techniques Part - II

Lecture 5: GPs and Streaming regression

An introduction to Sequential Monte Carlo

Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment

Lecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu

UNCERTAINTY QUANTIFICATION OF AIRFOIL DESIGN FOR ADVANCED PROPELLER MODELS

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Expectation Propagation Algorithm

MODULE -4 BAYEIAN LEARNING

Multi-fidelity co-kriging models

Bayesian analysis in nuclear physics

Concepts and Practice of Verification, Validation, and Uncertainty Quantification

Transcription:

Gaussian Processes for Computer Experiments Jeremy Oakley School of Mathematics and Statistics, University of Sheffield www.jeremy-oakley.staff.shef.ac.uk 1 / 43

Computer models Computer model represented by function y = f(x). f usually not available in closed form. 2 / 43

Computer models Computer model represented by function y = f(x). f usually not available in closed form. f constructed from modeller s understanding of the process. A substantive or law-driven model rather than an empirical/statistical model. 2 / 43

Computer models Computer model represented by function y = f(x). f usually not available in closed form. f constructed from modeller s understanding of the process. A substantive or law-driven model rather than an empirical/statistical model. There may be no physical (x, y) data. 2 / 43

Computer models Computer model represented by function y = f(x). f usually not available in closed form. f constructed from modeller s understanding of the process. A substantive or law-driven model rather than an empirical/statistical model. There may be no physical (x, y) data. f may be deterministic. 2 / 43

Computer models Computer model represented by function y = f(x). f usually not available in closed form. f constructed from modeller s understanding of the process. A substantive or law-driven model rather than an empirical/statistical model. There may be no physical (x, y) data. f may be deterministic. Computer experiment: evaluating f at difference choices of x A model run : evaluating f at a single choice of x. 2 / 43

Things we want to do with computer models 1 Optimisation: finding input values to produce desired output values 3 / 43

Things we want to do with computer models 1 Optimisation: finding input values to produce desired output values 2 Calibration/inverse problems: finding input values so that output values match observed data 3 / 43

Things we want to do with computer models 1 Optimisation: finding input values to produce desired output values 2 Calibration/inverse problems: finding input values so that output values match observed data 3 Uncertainty propagation. Suppose there is a true, uncertain input X. What is the distribution of Y = f(x)? 3 / 43

Things we want to do with computer models 1 Optimisation: finding input values to produce desired output values 2 Calibration/inverse problems: finding input values so that output values match observed data 3 Uncertainty propagation. Suppose there is a true, uncertain input X. What is the distribution of Y = f(x)? 4 (Global) sensitivity analysis. Continuing (3), if X = (X 1,..., X d ), how do elements of X contribute to uncertainty in Y? 3 / 43

Things we want to do with computer models 1 Optimisation: finding input values to produce desired output values 2 Calibration/inverse problems: finding input values so that output values match observed data 3 Uncertainty propagation. Suppose there is a true, uncertain input X. What is the distribution of Y = f(x)? 4 (Global) sensitivity analysis. Continuing (3), if X = (X 1,..., X d ), how do elements of X contribute to uncertainty in Y? New(ish) buzzword: Uncertainty Quantification (UQ) 3 / 43

Gaussian processes for computer models Computer models can be computationally expensive: takes long time to calculate f(x) for a single x. 4 / 43

Gaussian processes for computer models Computer models can be computationally expensive: takes long time to calculate f(x) for a single x. Given limited model runs y 1 = f(x 1 ),..., y n = f(x n ), wish to know y = f(x) for many other x values 4 / 43

Gaussian processes for computer models Computer models can be computationally expensive: takes long time to calculate f(x) for a single x. Given limited model runs y 1 = f(x 1 ),..., y n = f(x n ), wish to know y = f(x) for many other x values You ll have seen by now that GPs are good for this sort of thing! 4 / 43

Gaussian processes for computer models 14 12 10 f(x) 8 6 4 2 0 1 2 3 4 5 6 7 8 9 10 x 5 / 43

Computer experiments: some considerations 6 / 43

Computer experiments: some considerations f is often deterministic 6 / 43

Computer experiments: some considerations f is often deterministic Usually working with/aiming for small training datasets 6 / 43

Computer experiments: some considerations f is often deterministic Usually working with/aiming for small training datasets Can choose the training inputs: experimental design problem 6 / 43

Computer experiments: some considerations f is often deterministic Usually working with/aiming for small training datasets Can choose the training inputs: experimental design problem Can sometimes work with multi-level computer models, eg, a fast and coarse model and a slow and accurate model of the same system 6 / 43

Past and current research First(?) use of a Gaussian process for a computer model: Sacks, J., Welch, W.J., Mitchell, T.J. and Wynn, H.P. (1989) Design and analysis of computer experiments. Statistical Science, 4, 409-435. 7 / 43

Past and current research First(?) use of a Gaussian process for a computer model: Sacks, J., Welch, W.J., Mitchell, T.J. and Wynn, H.P. (1989) Design and analysis of computer experiments. Statistical Science, 4, 409-435. Bayesian version: Currin, C., Mitchell, T., Morris, M. and Ylvisaker, D. (1991). Bayesian prediction of deterministic functions, with applications to the design and analysis of computer experiments Journal of the American Statistical Association, 86, 953-963. 7 / 43

Past and current research First(?) use of a Gaussian process for a computer model: Sacks, J., Welch, W.J., Mitchell, T.J. and Wynn, H.P. (1989) Design and analysis of computer experiments. Statistical Science, 4, 409-435. Bayesian version: Currin, C., Mitchell, T., Morris, M. and Ylvisaker, D. (1991). Bayesian prediction of deterministic functions, with applications to the design and analysis of computer experiments Journal of the American Statistical Association, 86, 953-963. GPs for optimisation: Jones, D.R., Schonlau, M. and Welch, W. J. (1998). Efficient Global Optimization of Expensive Black-Box Functions, Journal of Global Optimization, 13, 455-492. 7 / 43

Past and current research First(?) use of a Gaussian process for a computer model: Sacks, J., Welch, W.J., Mitchell, T.J. and Wynn, H.P. (1989) Design and analysis of computer experiments. Statistical Science, 4, 409-435. Bayesian version: Currin, C., Mitchell, T., Morris, M. and Ylvisaker, D. (1991). Bayesian prediction of deterministic functions, with applications to the design and analysis of computer experiments Journal of the American Statistical Association, 86, 953-963. GPs for optimisation: Jones, D.R., Schonlau, M. and Welch, W. J. (1998). Efficient Global Optimization of Expensive Black-Box Functions, Journal of Global Optimization, 13, 455-492. The MUCM community and toolkit www.mucm.ac.uk www.mucm.ac.uk/toolkit 7 / 43

Gaussian process emulators 8 / 43

Gaussian process emulators Various different names for the same thing: surrogate models, response surface models, meta-models, emulators 8 / 43

Gaussian process emulators Various different names for the same thing: surrogate models, response surface models, meta-models, emulators Emulator: a probability distribution for the function f 8 / 43

Gaussian process emulators Various different names for the same thing: surrogate models, response surface models, meta-models, emulators Emulator: a probability distribution for the function f Need more than an estimate; need to quantify uncertainty, as sample of training data may be small 8 / 43

Gaussian process emulators Various different names for the same thing: surrogate models, response surface models, meta-models, emulators Emulator: a probability distribution for the function f Need more than an estimate; need to quantify uncertainty, as sample of training data may be small Naturally handled within a Bayesian framework; no true distribution for f. We treat f as an uncertain function. 8 / 43

Gaussian process emulators Various different names for the same thing: surrogate models, response surface models, meta-models, emulators Emulator: a probability distribution for the function f Need more than an estimate; need to quantify uncertainty, as sample of training data may be small Naturally handled within a Bayesian framework; no true distribution for f. We treat f as an uncertain function. Emulator does not replace the model; arguably the term is redundant 8 / 43

Gaussian process emulators Various different names for the same thing: surrogate models, response surface models, meta-models, emulators Emulator: a probability distribution for the function f Need more than an estimate; need to quantify uncertainty, as sample of training data may be small Naturally handled within a Bayesian framework; no true distribution for f. We treat f as an uncertain function. Emulator does not replace the model; arguably the term is redundant In a computer experiment, may want to know f(x 1 ),..., f(x N ), but can only observe f(x 1 ),..., f(x n ), with n < N. 8 / 43

Gaussian process emulators Various different names for the same thing: surrogate models, response surface models, meta-models, emulators Emulator: a probability distribution for the function f Need more than an estimate; need to quantify uncertainty, as sample of training data may be small Naturally handled within a Bayesian framework; no true distribution for f. We treat f as an uncertain function. Emulator does not replace the model; arguably the term is redundant In a computer experiment, may want to know f(x 1 ),..., f(x N ), but can only observe f(x 1 ),..., f(x n ), with n < N. As part of the analysis, we work with p{f(x n+1 ),..., f(x N ) f(x 1 ),..., f(x n )}, which we get from the emulator. 8 / 43

GP emulator modelling choices 9 / 43

GP emulator modelling choices Convention is to specify the GP model hierarchically, with separate mean and covariance functions. f(x) = h(x) T β + Z(x) 9 / 43

GP emulator modelling choices Convention is to specify the GP model hierarchically, with separate mean and covariance functions. f(x) = h(x) T β + Z(x) h(x) T β: the mean function, eg, for scalar x, could have h(x) T β = β 0 + β 1 x. 9 / 43

GP emulator modelling choices Convention is to specify the GP model hierarchically, with separate mean and covariance functions. f(x) = h(x) T β + Z(x) h(x) T β: the mean function, eg, for scalar x, could have h(x) T β = β 0 + β 1 x. Z(.): a zero-mean GP (usually stationary), sometimes called the residual process 9 / 43

GP emulator modelling choices Convention is to specify the GP model hierarchically, with separate mean and covariance functions. f(x) = h(x) T β + Z(x) h(x) T β: the mean function, eg, for scalar x, could have h(x) T β = β 0 + β 1 x. Z(.): a zero-mean GP (usually stationary), sometimes called the residual process Can think of the role of Z(.) as improving a parametric model 9 / 43

GP emulator modelling choices Convention is to specify the GP model hierarchically, with separate mean and covariance functions. f(x) = h(x) T β + Z(x) h(x) T β: the mean function, eg, for scalar x, could have h(x) T β = β 0 + β 1 x. Z(.): a zero-mean GP (usually stationary), sometimes called the residual process Can think of the role of Z(.) as improving a parametric model One aim of the mean function is to ensure stationary GP is adequate 9 / 43

Have and E[Z(x)] = 0 Cov(Z(x), Z(x ) σ 2, θ) = σ 2 c θ (x, x ) 10 / 43

Have E[Z(x)] = 0 and Cov(Z(x), Z(x ) σ 2, θ) = σ 2 c θ (x, x ) with, eg, for x = (x 1,..., x d ) { c(x, x ) = exp d ( xi x ) } 2 i i=1 θ i 10 / 43

Have E[Z(x)] = 0 and Cov(Z(x), Z(x ) σ 2, θ) = σ 2 c θ (x, x ) with, eg, for x = (x 1,..., x d ) { c(x, x ) = exp Commonly assume d ( xi x ) } 2 i i=1 p(β, σ 2 ) 1 σ 2 θ i 10 / 43

Have E[Z(x)] = 0 and Cov(Z(x), Z(x ) σ 2, θ) = σ 2 c θ (x, x ) with, eg, for x = (x 1,..., x d ) { c(x, x ) = exp Commonly assume d ( xi x ) } 2 i i=1 p(β, σ 2 ) 1 σ 2 θ i Given training data (and θ), posterior distribution of f is a t-process (β and σ 2 integrated out of posterior analytically) 10 / 43

Equivalent ways to parameterise a GP 11 / 43

Equivalent ways to parameterise a GP Consider, for scalar input x, f(x) = xβ + Z(x) with β σ 2 N(0, vσ 2 ) (with v chosen), and Z(x) a zero mean GP as before. 11 / 43

Equivalent ways to parameterise a GP Consider, for scalar input x, f(x) = xβ + Z(x) with β σ 2 N(0, vσ 2 ) (with v chosen), and Z(x) a zero mean GP as before. Then E[f(x)] = 0 11 / 43

Equivalent ways to parameterise a GP Consider, for scalar input x, f(x) = xβ + Z(x) with β σ 2 N(0, vσ 2 ) (with v chosen), and Z(x) a zero mean GP as before. Then E[f(x)] = 0 and Cov(f(x 1 ), f(x 2 ) σ 2, θ) = vσ 2 x 1 x 2 + σ 2 c θ (x 1, x 2 ) (1) 11 / 43

Equivalent ways to parameterise a GP Consider, for scalar input x, f(x) = xβ + Z(x) with β σ 2 N(0, vσ 2 ) (with v chosen), and Z(x) a zero mean GP as before. Then E[f(x)] = 0 and Cov(f(x 1 ), f(x 2 ) σ 2, θ) = vσ 2 x 1 x 2 + σ 2 c θ (x 1, x 2 ) (1) If we specify a zero mean GP with covariance function (1), for fixed v, model is equivalent to the hierarchical GP with the mean function xβ. 11 / 43

Choice of mean function 12 / 43

Choice of mean function Can be important if training data set small; gaps between points can be large. 12 / 43

Choice of mean function Can be important if training data set small; gaps between points can be large. Usually choose simple linear form. If x = (x 1,..., x d ), h(x) T β = β 0 + β 1 x 1 +... + β d x d. 12 / 43

Choice of mean function Can be important if training data set small; gaps between points can be large. Usually choose simple linear form. If x = (x 1,..., x d ), h(x) T β = β 0 + β 1 x 1 +... + β d x d. Some claim h(x) T β = β 0 constant form works better... 12 / 43

Choice of mean function Can be important if training data set small; gaps between points can be large. Usually choose simple linear form. If x = (x 1,..., x d ), h(x) T β = β 0 + β 1 x 1 +... + β d x d. Some claim h(x) T β = β 0 constant form works better......others claim better to include higher order polynomial terms (quadratics, interactions) 12 / 43

Choice of mean function Can be important if training data set small; gaps between points can be large. Usually choose simple linear form. If x = (x 1,..., x d ), h(x) T β = β 0 + β 1 x 1 +... + β d x d. Some claim h(x) T β = β 0 constant form works better......others claim better to include higher order polynomial terms (quadratics, interactions) For multi level model case, fast model can be used as a prior mean for the slow model 12 / 43

Choice of mean function Can be important if training data set small; gaps between points can be large. Usually choose simple linear form. If x = (x 1,..., x d ), h(x) T β = β 0 + β 1 x 1 +... + β d x d. Some claim h(x) T β = β 0 constant form works better......others claim better to include higher order polynomial terms (quadratics, interactions) For multi level model case, fast model can be used as a prior mean for the slow model Using a second GP (with noise) for the mean can help deal with nonstationarity: Ba, S. and Joseph, R. V. (2012). Composite Gaussian process models for emulating expensive functions. The Annals of Applied Statistics 6, 1838-1860 12 / 43

Estimating the correlation function parameters x = (x 1,..., x d ). { c(x, x ) = exp d ( xi x ) } 2 i i=1 θ i 13 / 43

Estimating the correlation function parameters x = (x 1,..., x d ). { c(x, x ) = exp d ( xi x ) } 2 i i=1 θ i Maximum likelihood probably most popular 13 / 43

Estimating the correlation function parameters x = (x 1,..., x d ). { c(x, x ) = exp d ( xi x ) } 2 i i=1 θ i Maximum likelihood probably most popular Some authors do full Bayes using MCMC 13 / 43

Estimating the correlation function parameters x = (x 1,..., x d ). { c(x, x ) = exp d ( xi x ) } 2 i i=1 θ i Maximum likelihood probably most popular Some authors do full Bayes using MCMC Others fix the correlation parameters, and include more polynomial terms in the mean 13 / 43

Estimating the correlation function parameters x = (x 1,..., x d ). { c(x, x ) = exp d ( xi x ) } 2 i i=1 θ i Maximum likelihood probably most popular Some authors do full Bayes using MCMC Others fix the correlation parameters, and include more polynomial terms in the mean Importance sampling: 13 / 43

Estimating the correlation function parameters x = (x 1,..., x d ). { c(x, x ) = exp d ( xi x ) } 2 i i=1 θ i Maximum likelihood probably most popular Some authors do full Bayes using MCMC Others fix the correlation parameters, and include more polynomial terms in the mean Importance sampling: Nagy, B., Loeppky, J. L. and Welch, W. J. (2007). Fast Bayesian inference for Gaussian process models. 13 / 43

Initial Gibbs sampler can be helpful for starting an optimiser. Example: 18 input climate model log likelihood -390-380 -370-360 -350-340 -330 0 200 400 600 800 1000 iteration 14 / 43

Experimental design 15 / 43

Experimental design Often find some inputs uninfluential; product designs can be wasteful 15 / 43

Experimental design Often find some inputs uninfluential; product designs can be wasteful x2 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 x1 15 / 43

Experimental design Often find some inputs uninfluential; product designs can be wasteful x2 0.0 0.2 0.4 0.6 0.8 1.0 x2 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 x1 0.0 0.2 0.4 0.6 0.8 1.0 x1 Space filling designs such as (maximin) Latin Hypercube samples popular 15 / 43

Experimental design Sobol sequences also space-filling. Easier for sequential design, but some projections poor. D[, 18] 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 D[, 11] 16 / 43

Experimental design Sobol sequences also space-filling. Easier for sequential design, but some projections poor. D[, 18] 0.0 0.2 0.4 0.6 0.8 1.0 D[, 18] 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 D[, 11] 0.0 0.2 0.4 0.6 0.8 1.0 D[, 10] 16 / 43

Uncertainty propagation Suppose that there is a true, uncertain value of the input X We specify a probability distribution for X 17 / 43

Uncertainty propagation Suppose that there is a true, uncertain value of the input X We specify a probability distribution for X What is our uncertainty about Y = f(x)? 17 / 43

Uncertainty propagation Suppose that there is a true, uncertain value of the input X We specify a probability distribution for X What is our uncertainty about Y = f(x)? We consider inference for the uncertainty distribution p(y f), itself uncertain. (Distinct from p(y )). 17 / 43

Uncertainty propagation Suppose that there is a true, uncertain value of the input X We specify a probability distribution for X What is our uncertainty about Y = f(x)? We consider inference for the uncertainty distribution p(y f), itself uncertain. (Distinct from p(y )). Example: Suppose X N(0, 1), and either f(x) = 1 + x, or f(x) = x 2 (we don t know which). 17 / 43

Uncertainty propagation Suppose that there is a true, uncertain value of the input X We specify a probability distribution for X What is our uncertainty about Y = f(x)? We consider inference for the uncertainty distribution p(y f), itself uncertain. (Distinct from p(y )). Example: Suppose X N(0, 1), and either f(x) = 1 + x, or f(x) = x 2 (we don t know which). If f(x) = 1 + x, then Y = f(x) f has the N(1, 1) distribution 17 / 43

Uncertainty propagation Suppose that there is a true, uncertain value of the input X We specify a probability distribution for X What is our uncertainty about Y = f(x)? We consider inference for the uncertainty distribution p(y f), itself uncertain. (Distinct from p(y )). Example: Suppose X N(0, 1), and either f(x) = 1 + x, or f(x) = x 2 (we don t know which). If f(x) = 1 + x, then Y = f(x) f has the N(1, 1) distribution If f(x) = x 2, then Y = f(x) f has the χ 2 1 distribution 17 / 43

Uncertainty propagation Suppose that there is a true, uncertain value of the input X We specify a probability distribution for X What is our uncertainty about Y = f(x)? We consider inference for the uncertainty distribution p(y f), itself uncertain. (Distinct from p(y )). Example: Suppose X N(0, 1), and either f(x) = 1 + x, or f(x) = x 2 (we don t know which). If f(x) = 1 + x, then Y = f(x) f has the N(1, 1) distribution If f(x) = x 2, then Y = f(x) f has the χ 2 1 distribution As f is uncertain, distribution of Y f is uncertain 17 / 43

Uncertainty propagation Suppose that there is a true, uncertain value of the input X We specify a probability distribution for X What is our uncertainty about Y = f(x)? We consider inference for the uncertainty distribution p(y f), itself uncertain. (Distinct from p(y )). Example: Suppose X N(0, 1), and either f(x) = 1 + x, or f(x) = x 2 (we don t know which). If f(x) = 1 + x, then Y = f(x) f has the N(1, 1) distribution If f(x) = x 2, then Y = f(x) f has the χ 2 1 distribution As f is uncertain, distribution of Y f is uncertain By considering uncertainty about p(y f), we can consider how inferences might change given more model runs 17 / 43

Consider inference for the uncertain mean E(Y f) = f(x)p X (x)dx, 18 / 43

Consider inference for the uncertain mean E(Y f) = f(x)p X (x)dx, Normally distributed if f is a GP (O Hagan, A. (1991) Bayes-Hermite quadrature). 18 / 43

Consider inference for the uncertain mean E(Y f) = f(x)p X (x)dx, Normally distributed if f is a GP (O Hagan, A. (1991) Bayes-Hermite quadrature). Can get closed form expression for mean and variance of E(Y f) if Multivariate normal distribution for X Exponentiated quadratic covariance function Mean function polynomial in X 18 / 43

Consider inference for the uncertain mean E(Y f) = f(x)p X (x)dx, Normally distributed if f is a GP (O Hagan, A. (1991) Bayes-Hermite quadrature). Can get closed form expression for mean and variance of E(Y f) if Multivariate normal distribution for X Exponentiated quadratic covariance function Mean function polynomial in X Haylock, R. G. and O Hagan, A. (1996). On inference for outputs of computationally expensive algorithms with uncertainty on the inputs, in Bayesian Statistics 5, edited by Bernardo, J. M., Berger, J. O., Dawid, A. P. and Smith, A. F. M., pp. 629-637, Oxford: University Press. 18 / 43

Exploring the uncertainty distribution with emulators: example Oakley, J. (2004). Estimating percentiles of computer code outputs. Journal of the Royal Statistical Society, Series C, 53, 83-93. 19 / 43

Exploring the uncertainty distribution with emulators: example Oakley, J. (2004). Estimating percentiles of computer code outputs. Journal of the Royal Statistical Society, Series C, 53, 83-93. During storms, excess sewage can spill over into rivers and streams 19 / 43

Exploring the uncertainty distribution with emulators: example Oakley, J. (2004). Estimating percentiles of computer code outputs. Journal of the Royal Statistical Society, Series C, 53, 83-93. During storms, excess sewage can spill over into rivers and streams A model y = f(x) (the SIMPOL model) estimates storage volume y required to meet certain environmental standards. 19 / 43

Exploring the uncertainty distribution with emulators: example Oakley, J. (2004). Estimating percentiles of computer code outputs. Journal of the Royal Statistical Society, Series C, 53, 83-93. During storms, excess sewage can spill over into rivers and streams A model y = f(x) (the SIMPOL model) estimates storage volume y required to meet certain environmental standards. Uncertainty in true input values X induces uncertainty in the output Y = f(x); WRc wish to know the 95th percentile of Y f 19 / 43

Exploring the uncertainty distribution with emulators: example Oakley, J. (2004). Estimating percentiles of computer code outputs. Journal of the Royal Statistical Society, Series C, 53, 83-93. During storms, excess sewage can spill over into rivers and streams A model y = f(x) (the SIMPOL model) estimates storage volume y required to meet certain environmental standards. Uncertainty in true input values X induces uncertainty in the output Y = f(x); WRc wish to know the 95th percentile of Y f Can we estimate this percentile efficiently based on a small number of model runs? 19 / 43

A sequential approach 20 / 43

A sequential approach 1 Run the model a small number times, fit emulator to get distribution of f 20 / 43

A sequential approach 1 Run the model a small number times, fit emulator to get distribution of f 2 Simulate approximate realisations of f. For each realisation, use Monte Carlo to estimate 95th percentile, noting the input value. 20 / 43

A sequential approach 1 Run the model a small number times, fit emulator to get distribution of f 2 Simulate approximate realisations of f. For each realisation, use Monte Carlo to estimate 95th percentile, noting the input value. 3 From (2), identify region of input space R that produces large output values 20 / 43

A sequential approach 1 Run the model a small number times, fit emulator to get distribution of f 2 Simulate approximate realisations of f. For each realisation, use Monte Carlo to estimate 95th percentile, noting the input value. 3 From (2), identify region of input space R that produces large output values 4 Run original more times, concentrating input choices in R 20 / 43

A sequential approach 1 Run the model a small number times, fit emulator to get distribution of f 2 Simulate approximate realisations of f. For each realisation, use Monte Carlo to estimate 95th percentile, noting the input value. 3 From (2), identify region of input space R that produces large output values 4 Run original more times, concentrating input choices in R 5 Update the emulator to reduce uncertainty about f in input region of interest 20 / 43

A sequential approach 1 Run the model a small number times, fit emulator to get distribution of f 2 Simulate approximate realisations of f. For each realisation, use Monte Carlo to estimate 95th percentile, noting the input value. 3 From (2), identify region of input space R that produces large output values 4 Run original more times, concentrating input choices in R 5 Update the emulator to reduce uncertainty about f in input region of interest SIMPOL test example, 4 uncertain inputs, 16 runs used for step 1, 8 runs used for step 3. 20 / 43

A sequential approach 1 Run the model a small number times, fit emulator to get distribution of f 2 Simulate approximate realisations of f. For each realisation, use Monte Carlo to estimate 95th percentile, noting the input value. 3 From (2), identify region of input space R that produces large output values 4 Run original more times, concentrating input choices in R 5 Update the emulator to reduce uncertainty about f in input region of interest SIMPOL test example, 4 uncertain inputs, 16 runs used for step 1, 8 runs used for step 3. Comparison with large Monte Carlo samples. 20 / 43

0.80 0.85 0.90 0.95 1.00 0.80 0.85 0.90 0.95 1.00 Bayes, n=24 1000 1050 1100 1150 1200 1250 1300 output Monte Carlo, n=500 1000 1050 1100 1150 1200 1250 1300 output 0.80 0.85 0.90 0.95 1.00 0.80 0.85 0.90 0.95 1.00 Monte Carlo, n=1000 1000 1050 1100 1150 1200 1250 1300 output Monte Carlo, n=100 1000 1050 1100 1150 1200 1250 1300 output 21 / 43

Diagnostics and validation 22 / 43

Diagnostics and validation Bastos, L. S. and O Hagan, A. (2009). Diagnostics for Gaussian process emulators. Technometrics 51, 425-438. 22 / 43

Diagnostics and validation Bastos, L. S. and O Hagan, A. (2009). Diagnostics for Gaussian process emulators. Technometrics 51, 425-438. Motivating example: We have a model y = f(x), and wish to consider uncertainty in Y = f(x), for uncertain X with distribution G 22 / 43

Diagnostics and validation Bastos, L. S. and O Hagan, A. (2009). Diagnostics for Gaussian process emulators. Technometrics 51, 425-438. Motivating example: We have a model y = f(x), and wish to consider uncertainty in Y = f(x), for uncertain X with distribution G Monte Carlo approach: sample x 1,..., x N from G, evaluate y 1 = f(x 1 ),..., y n = f(x N ) to obtain sample from Y. 22 / 43

Diagnostics and validation Bastos, L. S. and O Hagan, A. (2009). Diagnostics for Gaussian process emulators. Technometrics 51, 425-438. Motivating example: We have a model y = f(x), and wish to consider uncertainty in Y = f(x), for uncertain X with distribution G Monte Carlo approach: sample x 1,..., x N from G, evaluate y 1 = f(x 1 ),..., y n = f(x N ) to obtain sample from Y. Can use Monte Carlo sample to estimate E(Y f), P (Y y f) etc., with reliable associated standard errors for large N. 22 / 43

Diagnostics and validation Bastos, L. S. and O Hagan, A. (2009). Diagnostics for Gaussian process emulators. Technometrics 51, 425-438. Motivating example: We have a model y = f(x), and wish to consider uncertainty in Y = f(x), for uncertain X with distribution G Monte Carlo approach: sample x 1,..., x N from G, evaluate y 1 = f(x 1 ),..., y n = f(x N ) to obtain sample from Y. Can use Monte Carlo sample to estimate E(Y f), P (Y y f) etc., with reliable associated standard errors for large N. Model user limited to small number of runs, so do uncertainty analysis with an emulator. 22 / 43

Diagnostics and validation: motivating example Suppose emulator provides estimate Ê(Y f) = 50, with 95% credible interval (48,52) and ˆP (Y < 70 f) = 0.8, with 95% credible interval (0.77,0.83) 23 / 43

Diagnostics and validation: motivating example Suppose emulator provides estimate Ê(Y f) = 50, with 95% credible interval (48,52) and ˆP (Y < 70 f) = 0.8, with 95% credible interval (0.77,0.83) Intervals show how estimates might change given further runs of the model 23 / 43

Diagnostics and validation: motivating example Suppose emulator provides estimate Ê(Y f) = 50, with 95% credible interval (48,52) and ˆP (Y < 70 f) = 0.8, with 95% credible interval (0.77,0.83) Intervals show how estimates might change given further runs of the model Can we really be 95% sure E(Y f) lies in (48,52)? 23 / 43

Diagnostics and validation: motivating example Suppose emulator provides estimate Ê(Y f) = 50, with 95% credible interval (48,52) and ˆP (Y < 70 f) = 0.8, with 95% credible interval (0.77,0.83) Intervals show how estimates might change given further runs of the model Can we really be 95% sure E(Y f) lies in (48,52)? Will depend on whether GP emulator is an appropriate statistical model of the computer model f. 23 / 43

Diagnostics and validation: motivating example Suppose emulator provides estimate Ê(Y f) = 50, with 95% credible interval (48,52) and ˆP (Y < 70 f) = 0.8, with 95% credible interval (0.77,0.83) Intervals show how estimates might change given further runs of the model Can we really be 95% sure E(Y f) lies in (48,52)? Will depend on whether GP emulator is an appropriate statistical model of the computer model f. Overconfident predictions may lead to poor decisions 23 / 43

Diagnostics and validation: motivating example Suppose emulator provides estimate Ê(Y f) = 50, with 95% credible interval (48,52) and ˆP (Y < 70 f) = 0.8, with 95% credible interval (0.77,0.83) Intervals show how estimates might change given further runs of the model Can we really be 95% sure E(Y f) lies in (48,52)? Will depend on whether GP emulator is an appropriate statistical model of the computer model f. Overconfident predictions may lead to poor decisions Underconfident predictions may devalue the computer model analysis 23 / 43

Poorly performing emulators 8 3 6 2 4 1 η(x) 2 0 2 η(x) 0 1 4 2 6 3 8 1 0 1 x 4 1 0 1 x 24 / 43

Traditional approach: leave one out CV Remove each data point in turn, predict using remaining training data 25 / 43

Traditional approach: leave one out CV Remove each data point in turn, predict using remaining training data Plot prediction (emulator mean), with error bars, against true value Plot standardised prediction errors against predicted values 25 / 43

Traditional approach: leave one out CV Remove each data point in turn, predict using remaining training data Plot prediction (emulator mean), with error bars, against true value Plot standardised prediction errors against predicted values Useful, can identify a bad emulator, but ignores correlation in prediction errors hard to judge if uncertainty in emulator is appropriate 25 / 43

The diagnostic strategy Run simulator to obtain training data y. Build emulator given the training data. 4 3 2 1 η(x) 0 1 2 3 4 5 2 1.5 1 0.5 0 0.5 1 1.5 2 x 26 / 43

The diagnostic strategy Select second set validation inputs. Determine emulator predictions and uncertainty at these inputs. 4 3 2 1 η(x) 0 1 2 3 4 5 2 1.5 1 0.5 0 0.5 1 1.5 2 x 27 / 43

The diagnostic strategy Run simulator to obtain validation data y. Observe emulator prediction errors, relative to emulator uncertainty. 4 3 2 1 η(x) 0 1 2 3 4 5 2 1.5 1 0.5 0 0.5 1 1.5 2 x 28 / 43

Diagnostics with the validation data (1) Can combine prediction errors into a single measure of fit using a Mahalanobis distance D MD (y ) = {y E(y y)} T V ar(y y) 1 {y E(y y)}, How do we know if prediction errors are too large or too small? 29 / 43

Diagnostics with the validation data (1) Can combine prediction errors into a single measure of fit using a Mahalanobis distance D MD (y ) = {y E(y y)} T V ar(y y) 1 {y E(y y)}, How do we know if prediction errors are too large or too small? The Mahalanobis distance diagnostic If GP emulator is correct, D MD (y ) has an F distribution, so compare observed D MD (y ) with this distribution to test hypothesis of a good emulator. 29 / 43

Diagnostics with the validation data (2) Prediction errors in the validation data are correlated Transform errors using a pivoted Cholesky decomposition of V ar(y y). 30 / 43

Diagnostics with the validation data (2) Prediction errors in the validation data are correlated Transform errors using a pivoted Cholesky decomposition of V ar(y y). This gives us Sequence of independent errors, each approx N(0, 1) for a correct emulator 30 / 43

Diagnostics with the validation data (2) Prediction errors in the validation data are correlated Transform errors using a pivoted Cholesky decomposition of V ar(y y). This gives us Sequence of independent errors, each approx N(0, 1) for a correct emulator Early points in sequence correspond to most uncertain validation points - how does emulator behave in regions distant to the training data? 30 / 43

Diagnostics with the validation data (2) Prediction errors in the validation data are correlated Transform errors using a pivoted Cholesky decomposition of V ar(y y). This gives us Sequence of independent errors, each approx N(0, 1) for a correct emulator Early points in sequence correspond to most uncertain validation points - how does emulator behave in regions distant to the training data? Latter points in sequence correspond to validation points close to training data, or other validation points - are we describing the smoothness of the simulator appropriately? 30 / 43

The pivoted Cholesky decomposition diagnostic 8 3 η(x) 6 4 2 0 2 4 6 transformed error 2 1 0 1 2 8 2 0 2 x 3 0 5 10 15 20 index 31 / 43

Example: 18 input climate model, constant mean function Emulator means and 95% intervals Normal Q-Q Plot emulator prediction 12 14 16 18 Sample Quantiles -4-2 0 2 4 13 14 15 16 17 18 19 simulator output -2-1 0 1 2 Theoretical Quantiles Pivoted Cholesky prediction errors Mahalanobis test Pivoted Cholesky residual -4-2 0 2 4 Density 0.0 0.5 1.0 1.5 2.0 0 20 40 60 80 100 Pivot index 0.0 0.5 1.0 1.5 2.0 2.5 3.0 x 32 / 43

Example: 18 input climate model, linear mean function Emulator means and 95% intervals Normal Q-Q Plot emulator prediction 12 14 16 18 Sample Quantiles -3-2 -1 0 1 2 3 13 14 15 16 17 18 19 simulator output -2-1 0 1 2 Theoretical Quantiles Pivoted Cholesky prediction errors Mahalanobis test Pivoted Cholesky residual -3-2 -1 0 1 2 3 Density 0.0 0.5 1.0 1.5 2.0 0 20 40 60 80 100 Pivot index 0.0 0.5 1.0 1.5 x 33 / 43

Sensitivity analysis of model outputs 34 / 43

Sensitivity analysis of model outputs Local sensitivity analysis: calculating partial derivatives of f 34 / 43

Sensitivity analysis of model outputs Local sensitivity analysis: calculating partial derivatives of f Global or probabilistic sensitivity analysis: which elements in X = {X 1,..., X d } are most responsible for the uncertainty in Y = f(x)? 34 / 43

Sensitivity analysis of model outputs Local sensitivity analysis: calculating partial derivatives of f Global or probabilistic sensitivity analysis: which elements in X = {X 1,..., X d } are most responsible for the uncertainty in Y = f(x)? Variance based approach: investigate how inputs contribute to V ar(y ) (given f). Saltelli, A., Ratto, M., Andres, T., Campolongo, F., J.Cariboni, Gatelli, D., Saisana, M. and Tarantola, S. (2008). Global Sensitivity Analysis: The Primer, New York: Wiley. 34 / 43

Variance-based global sensitivity analysis Main effect variance: V ar Xi {E(Y X i )} 35 / 43

Variance-based global sensitivity analysis Main effect variance: V ar Xi {E(Y X i )} The expected reduction in variance if value of X i is learnt, because V ar Xi {E(Y X i )} = V ar(y ) E Xi {V ar(y X i )} 35 / 43

Variance-based global sensitivity analysis Main effect variance: V ar Xi {E(Y X i )} The expected reduction in variance if value of X i is learnt, because V ar Xi {E(Y X i )} = V ar(y ) E Xi {V ar(y X i )} If inputs independent, can decompose V ar(y ) into main effect variances and interaction variances, similar to ANOVA. 35 / 43

Variance-based global sensitivity analysis Main effect variance: V ar Xi {E(Y X i )} The expected reduction in variance if value of X i is learnt, because V ar Xi {E(Y X i )} = V ar(y ) E Xi {V ar(y X i )} If inputs independent, can decompose V ar(y ) into main effect variances and interaction variances, similar to ANOVA. Depends on specification of p(x). 35 / 43

Variance-based global sensitivity analysis Main effect variance: V ar Xi {E(Y X i )} The expected reduction in variance if value of X i is learnt, because V ar Xi {E(Y X i )} = V ar(y ) E Xi {V ar(y X i )} If inputs independent, can decompose V ar(y ) into main effect variances and interaction variances, similar to ANOVA. Depends on specification of p(x). Can measure input importance in different ways, eg decision-theoretic approaches. 35 / 43

Main effect plots and variances 1 0.9 E(Y X i ) E(Y) p(x i ) 1 0.9 E(Y X j ) E(Y) p(x j ) 0.8 0.8 0.7 0.7 0.6 0.6 0.5 0.5 0.4 0.4 0.3 0.3 0.2 0.2 0.1 0.1 0 5 0 5 X i 0 5 0 5 X j 36 / 43

Computation with emulators Bayesian quadrature again. Now want E(Y X i, f) = f(x i, X i )p X i (x i X i )dx i, 37 / 43

Computation with emulators Bayesian quadrature again. Now want E(Y X i, f) = f(x i, X i )p X i (x i X i )dx i, Again, normally distributed with GP model for f. Can obtain mean and variance analytically for certain GP modelling choices and normally distributed inputs. 37 / 43

Computation with emulators Bayesian quadrature again. Now want E(Y X i, f) = f(x i, X i )p X i (x i X i )dx i, Again, normally distributed with GP model for f. Can obtain mean and variance analytically for certain GP modelling choices and normally distributed inputs. Oakley, J. E. and O Hagan, A. (2004). Probabilistic sensitivity of complex models: a Bayesian approach, J. Roy. Statist. Soc. Ser. B, 66: 751-769. 37 / 43

Variance-based SA for an epidemic model Acknowledgement: JP Gosling and others. See www.mucm.ac.uk/toolkit Case Study 1 % of total output variance 30 25 20 15 10 5 main effect interactions 0 0 5 10 15 20 input number 38 / 43

Calibration and inverse problems 39 / 43

Calibration and inverse problems Kennedy, M. and O Hagan, A. (2001). Bayesian calibration of computer models (with discussion). Journal of the Royal Statistical Society, Series B. 63, 425-464. 39 / 43

Calibration and inverse problems Kennedy, M. and O Hagan, A. (2001). Bayesian calibration of computer models (with discussion). Journal of the Royal Statistical Society, Series B. 63, 425-464. Example. A Gaussian plume deposition model f(x, θ) predicts deposition of radionuclides at a location x following release of concentration θ from point source. θ is unknown. 39 / 43

Calibration and inverse problems Kennedy, M. and O Hagan, A. (2001). Bayesian calibration of computer models (with discussion). Journal of the Royal Statistical Society, Series B. 63, 425-464. Example. A Gaussian plume deposition model f(x, θ) predicts deposition of radionuclides at a location x following release of concentration θ from point source. θ is unknown. Measurements of the true deposition z(x) at a limited number of locations x available. 39 / 43

Calibration and inverse problems Kennedy, M. and O Hagan, A. (2001). Bayesian calibration of computer models (with discussion). Journal of the Royal Statistical Society, Series B. 63, 425-464. Example. A Gaussian plume deposition model f(x, θ) predicts deposition of radionuclides at a location x following release of concentration θ from point source. θ is unknown. Measurements of the true deposition z(x) at a limited number of locations x available. Aim: to predict deposition at other locations using both data and model. 39 / 43

Calibration and inverse problems Kennedy, M. and O Hagan, A. (2001). Bayesian calibration of computer models (with discussion). Journal of the Royal Statistical Society, Series B. 63, 425-464. Example. A Gaussian plume deposition model f(x, θ) predicts deposition of radionuclides at a location x following release of concentration θ from point source. θ is unknown. Measurements of the true deposition z(x) at a limited number of locations x available. Aim: to predict deposition at other locations using both data and model. What value of θ do we use? 39 / 43

The calibration model z(x i ) = ρf(x i, θ) + δ(x i ) + ɛ i δ(x i ) is the discrepancy (bias) between model output and reality. Modelled with another GP. 40 / 43

The calibration model z(x i ) = ρf(x i, θ) + δ(x i ) + ɛ i δ(x i ) is the discrepancy (bias) between model output and reality. Modelled with another GP. Need discrepancy 1 to quantify uncertainty properly; 40 / 43

The calibration model z(x i ) = ρf(x i, θ) + δ(x i ) + ɛ i δ(x i ) is the discrepancy (bias) between model output and reality. Modelled with another GP. Need discrepancy 1 to quantify uncertainty properly; 2 to learn about a true θ 40 / 43

The calibration model z(x i ) = ρf(x i, θ) + δ(x i ) + ɛ i δ(x i ) is the discrepancy (bias) between model output and reality. Modelled with another GP. Need discrepancy 1 to quantify uncertainty properly; 2 to learn about a true θ But hard to specify; prior is important 40 / 43

The calibration model z(x i ) = ρf(x i, θ) + δ(x i ) + ɛ i δ(x i ) is the discrepancy (bias) between model output and reality. Modelled with another GP. Need discrepancy 1 to quantify uncertainty properly; 2 to learn about a true θ But hard to specify; prior is important Doesn t always go down well with modellers! 40 / 43

The calibration model z(x i ) = ρf(x i, θ) + δ(x i ) + ɛ i δ(x i ) is the discrepancy (bias) between model output and reality. Modelled with another GP. Need discrepancy 1 to quantify uncertainty properly; 2 to learn about a true θ But hard to specify; prior is important Doesn t always go down well with modellers! I m horrified! You should be improving your models with better physics! 40 / 43

History matching Emphasis on ruling out implausible regions of input space, prior to calibration 41 / 43

History matching Emphasis on ruling out implausible regions of input space, prior to calibration 1 Start with training runs D 1 = {y 1 = f(x 1 ),..., y n = f(x n )} over some design region 2 Fit emulator to the training data 41 / 43

History matching Emphasis on ruling out implausible regions of input space, prior to calibration 1 Start with training runs D 1 = {y 1 = f(x 1 ),..., y n = f(x n )} over some design region 2 Fit emulator to the training data 3 (Conservatively), discard implausible regions: regions where the best observed fit is poor May still be fairly uncertain about f at this stage, given D 1 only 41 / 43

History matching Emphasis on ruling out implausible regions of input space, prior to calibration 1 Start with training runs D 1 = {y 1 = f(x 1 ),..., y n = f(x n )} over some design region 2 Fit emulator to the training data 3 (Conservatively), discard implausible regions: regions where the best observed fit is poor May still be fairly uncertain about f at this stage, given D 1 only 4 Get second training set D 2 = {y n+1 = f(x n+1 ),..., y n = f(x n+m )} over the plausible design region 41 / 43

History matching Emphasis on ruling out implausible regions of input space, prior to calibration 1 Start with training runs D 1 = {y 1 = f(x 1 ),..., y n = f(x n )} over some design region 2 Fit emulator to the training data 3 (Conservatively), discard implausible regions: regions where the best observed fit is poor May still be fairly uncertain about f at this stage, given D 1 only 4 Get second training set D 2 = {y n+1 = f(x n+1 ),..., y n = f(x n+m )} over the plausible design region 5 Now discard further implausible regions, with improved information about f where it is needed 41 / 43

History matching Emphasis on ruling out implausible regions of input space, prior to calibration 1 Start with training runs D 1 = {y 1 = f(x 1 ),..., y n = f(x n )} over some design region 2 Fit emulator to the training data 3 (Conservatively), discard implausible regions: regions where the best observed fit is poor May still be fairly uncertain about f at this stage, given D 1 only 4 Get second training set D 2 = {y n+1 = f(x n+1 ),..., y n = f(x n+m )} over the plausible design region 5 Now discard further implausible regions, with improved information about f where it is needed 6 Iterate as necessary 41 / 43

History matching Emphasis on ruling out implausible regions of input space, prior to calibration 1 Start with training runs D 1 = {y 1 = f(x 1 ),..., y n = f(x n )} over some design region 2 Fit emulator to the training data 3 (Conservatively), discard implausible regions: regions where the best observed fit is poor May still be fairly uncertain about f at this stage, given D 1 only 4 Get second training set D 2 = {y n+1 = f(x n+1 ),..., y n = f(x n+m )} over the plausible design region 5 Now discard further implausible regions, with improved information about f where it is needed 6 Iterate as necessary I Vernon, M Goldstein and R G Bower (2010), "Galaxy Formation: A Bayesian Uncertainty Analysis", Bayesian Analysis 05(04): 619 670 (with discussion) 41 / 43

History matching: oil reservoir simulator Acknowledgement: JP Gosling and others. See www.mucm.ac.uk/toolkit Case Study 3 42 / 43

And finally... If you like Sheffield, and you like GPs: Uncertainty in Computer Models 2014 Sheffield July 28-30 www.mucm.ac.uk/ucm2014.html Abstract submissions for talks and posters welcome 43 / 43