Bayesian optimal design for Gaussian process models

Size: px
Start display at page:

Download "Bayesian optimal design for Gaussian process models"

Transcription

1 Bayesian optimal design for Gaussian process models Maria Adamou and Dave Woods Southampton Statistical Sciences Research Institute University of Southampton 1

2 Outline Background and motivation Computer experiments Spatial data collection Physical experiments Bayesian optimal design for GP regression Decision-theoretic design for prediction and estimation Approximating the objective function Numerical examples 2

3 Background Experiments are used to investigate the impact of an intervention ( treatment ) on a process or system through its application to a number of objects ( units ) Design of experiments - the active selection of the treatments (settings of the controllable variables) to be applied taking account of multiple sources of uncertainty Usually, the aim of the experiment is to collect data to answer scientific questions through the estimation of a statistical model The design can be selected to maximise information gained for a given resource Information is typically measured with respect to uncertainty about the model and its parameters 3

4 Motivation We consider design of experiments from a Bayesian prospective For example, prior information is used to choose the factors to study, their possible values and plausible models relating factors to the response This information may be subjective, empirical or mechanistic A Bayesian approach allows uncertainty in these decisions to be incorporated into selection of an appropriate design for the experiment For some experiments, it is natural or even necessary to be Bayesian 4

5 Computional modelling Many physical and social processes can be approximated by computer codes which encapsulate mathematical descriptions For example, partial differential equations, solved using e.g. finite element methods Inputs x Computer code g(x) Outputs y A computer experiment is performed when g(x) is computationally expensive, and hence evaluation of g(x) for all x is infeasible Choose a design x 1,..., x n, evaluate g(x 1),..., g(x n) and fit a statistical model (emulator) to this data The most common emulator is the Gaussian Process 5

6 Ride optimisation Jaguar-Land Rover offers a range of customisations for each model Aim: use a computer model to ensure satisfactory ride characteristics of 10 vehicle variants 6

7 Spatial statistics: sensor placement Spatial data is collected in a broad range of applications e.g. environmental, climatology and agriculture prediction candidate optimal Monitoring spatial sulphate deposition in the eastern USA 122 candidate locations, 10 prediction points and 40 design points Gaussian process model is commonly employed Where should monitoring stations be sited for best prediction? Cressie (1993) 7

8 Physical experiments Choose treatment settings in the presence of e.g. spatial or temporal trends for example, field trials in agriculture, lab experiments in well-plates Assume random effects for nuisance factors (fields, plates,...) with covariance of observed response depending on distance between units 8

9 Gaussian process model 9

10 Gaussian process regression Bayesian Gaussian process models: nonparametric approach to modelling noisy data y(x) = g(x) + ε, ε N(0, σ 2 ε) (1) Assume a Gaussian process prior g(x) GP {µ(x), κ(x, x )} g(x), g(x ) joint Gaussian E{g(x)} = µ(x) = f T (x)β trend cov ( g(x) g(x ) ) = 1 k σ2 g ( λ (x, x ) k λ (x, x ) covariance ) 1 Rasmussen & Williams (2006) 10

11 Inference Updated uncertainty, incorporating the data, is given by the posterior distribution g y Assuming the Gaussian process prior, it is straightforward to derive the posterior conditional distribution where g(x) y, β, λ, σ 2 g, σ 2 ε N[ m(x), σ 2 gs 2 ] m(x) = f T (x)β + k T L 1 n (y F β), k = [k λ (x, x i)] n i=1 L n = [k λ (x i, x j) + τ 2 δ ij] n i,j=1, s 2 = 1 k T L 1 n k, and τ 2 = σ 2 ɛ /σ 2 g, δ ij = 1 if i = j, 0 otherwise 11

12 A simple example g(x) = sin(x) µ(x) = β 0 (constant) k λ (x, x ) exp{ λ 2 (x x ) 2 } Posterior uncertainty approaches 0 at the design points Prediction Prediction 95% intervals Prior mean sin(x) x Learning λ and τ 2 requires numerical methods Empirical Bayes estimate λ, τ 2 (e.g. as posterior mode) and plug-in Markov chain Monte Carlo or Sequential Monte Carlo Titsias et al. (2011); Gramacy & Polson (2011) 12

13 Design for learning and prediction using GPs Space-filling: Latin Hypercubes, maximin and minimax,... McKay et al. (1979). Johnson et al., (1990),... Prediction: kriging variance, usually assuming known λ, IMSPE, Bayesian approaches... O Hagan (1978), Sacks et al., (1989), Zhou & Stein (2006), Diggle & Lophaven (2006), Gorodetsky & Marzouk (2016), Leatherman et al. (2017),... Estimation: Entropy, D-optimality,... Sebastiani & Wynn (2000), Zhou & Stein (2005), Xia et al. (2006), Boukouvalas et al. (2014),... Sequential design: optimisation (Expected Improvement), prediction,... Jones et al. (1998), Wikle & Royle (2005), Gramacy & Lee (2009), Beck & Guillas (2016),... 13

14 Bayesian Optimal Design 14

15 Decision theoretic design Loss functions L(d(y), θ; ξ) defines the cost of decision d(y) when the true state of nature is θ, where data y is obtained from design ξ = {x 1,..., x n} Example loss functions include squared error loss negative log-score 0-1 loss Make the decision that minimises the expected posterior loss d (y) = argmin d D L (d(y), θ; ξ) π(θ y, ξ) dθ Θ d (y) - Bayes decision for data y 15

16 Decision theoretic design Bayes risk Ψ(ξ) = Y Θ L (d (y), θ; ξ) π(θ, y ξ) dθdy Loss evaluated at Bayes decision for each y Y π(θ, y ξ) is the joint distribution of model parameters and data A Bayesian optimal design is then given by ξ = argmin ξ Ξ c.f. Berger (1985 Ch.7 preposterior analysis) Ψ(ξ) Lindley (1972), Chaloner & Verdinelli (1995) 16

17 Prediction 17

18 Prediction For some experiments, the aim is accurate model-free prediction of the response e.g. spatial studies, computer experiments Bayesian nonparametrics Place a Gaussian process prior on the unknown response function Correlation function allows predictions to smoothly follow data Optimal design by minimising prediction uncertainty O Hagan (1978), Müller et al., (2004), Diggle & Lophaven (2006), Adamou (2014) 18

19 Space-filling designs Space-filling designs are most commonly applied to experiments tailored for prediction Don t rely on the functional form of the relationship between inputs and outputs Good coverage is important for prediction (strongly influenced by nearby points) 19

20 Space-filling designs For Gaussian process model with (i) constant mean function and (ii) known values of correlation parameters: maximin optimal designs are asymptotically D-optimal minimax optimal designs are asymptotically G-optimal Johnson et al., (1990) D-optimal design maximises the generalised variance of the parameter estimators G-optimal design minimises the maximum predictive variance Atkinson et al. (2007) 20

21 Bayesian optimal design for prediction For prediction of future observation y(x n+) at unobserved point x n+, we apply the square error loss function L(ỹ, y(x n+); ξ) = ỹ y(x n+) 2 Best prediction of y(x n+) is ŷ = E(y(x n+) y) (Bayes decision) The objective function to be minimised is Ψ(ξ) = L(ŷ, y(x n+); ξ)π(y(x n+) y; ξ)π(y; ξ) dy(x n+) dy dx n+ = E[ ŷ y(x n+) 2 y; ξ]π(y; ξ) dy dx n+ = var(y(x n+) y; ξ)π(y; ξ) dy dx n+ 21

22 Evaluating the objective function For conjugate priors on β and σ 2 g 1. Assume φ = (λ, τ 2 ) are known: Ψ(ξ; φ) = var(y(x n+) φ, y; ξ)π(y φ; ξ) dy dx n+ Tractable integral with respect to the data Common in the literature, e.g Diggle & Lophaven (2006), Zimmerman (2006) 2. Unknown φ: Ψ(ξ) = var(y(x n+) y; ξ)π(y φ; ξ)π(φ) dφ dy dx n+ The integrand is intractable Numerical methods are required 22

23 An approximation Decompose the objective function Ψ(ξ) = E φ y [var(y(x n+) φ, y; ξ)] π(y; ξ) dy dx n+ + var φ y [E(y(x n+) φ, y; ξ)] π(y; ξ) dy dx n+ = Ψ 1(ξ) + Ψ 2(ξ) The integrand in Ψ 1(ξ) is analytically tractable with respect to the data y To approximate Ψ 2(ξ) options include: Markov Chain Monte Carlo simulations and Monte Carlo integration which result in low bias, high variance and high computation cost Fixed Ψ 2(ξ) = 0 with potentially high bias, zero variance and zero computation cost 23

24 An approximation Set Ψ 2(ξ) = 0, and use the approximation Ψ(ξ) Ψ 1(ξ) = E φ y [var(y(x n+) φ, y; ξ)] π(y; ξ) dy dx n+ = var(y(x n+) φ, y; ξ)π(y φ; ξ)π(φ) dy dφ dx n+ S 2 (x n+; φ, ξ)π(φ) dφ dx n+ where S 2 (x n+; φ, ξ) is the generalised kriging variance at the point x n+, and the integral with respect φ can be approximated using quadrature Gen. kriging var. 24

25 Assessing the approximation Numerical sensitivity study Generated random designs, for various n and prior hyperparameters, and evaluated Ψ 1 and Ψ 2 (using simulation) Consistently found Ψ 2 << Ψ 1 (i.e. Ψ Ψ 1) Ranking of designs very similar under Ψ 1 and Ψ, esp. for small Ψ Theoretical insights Ψ 2 depends on π(φ y; ξ) π(y φ; ξ)π(φ) Under some assumptions, integrated likelihood π(y φ; ξ) is bounded by a polynomial in φ E.g. as (λ, τ 2 ) (0, 0), π(y φ; ξ) 0 at a faster rate In the literature Berger et al. (2001) Wu & Kaufman (2013): studied predictive performance wrt choice of prior, found var φ y [E(y(x n+) φ, y; ξ)] 0 Leatherman et. al (2017a, 2017b): Bayesian-IMSPE (frequentist derivation, non-informative prior) 25

26 An approximation Approximate integral in Ψ(ξ) with respect to the prediction points Fixed grid results in zero variance and high bias low computation cost Monte Carlo grid results in high variance and low (zero) bias high computation cost For special cases of correlation function, i.e. Gaussian correlation function the integral with respect to the prediction points is analytically tractable 26

27 Computer experiments example Computer model: simple simulator of a helical compression spring Wire diameter (mm) Spring index Spring coefficient spring.jpg Three variables; linear mean function f T (x)β = β k=1 β k x k k λ (x, x ) = 3 j=1 exp( λ j x j x j ) for x, x [ 1, 1] 3 Prior for β, σ 2 g is Normal Inverse Gamma Bayesian optimal designs with n = 10 runs for two different prior distributions Prior 1: τ 2 =0, λ 1 Unif(1, 3), λ 2 Unif(3, 5), λ 3 0 Prior 2: τ 2 =0.5, λ 1 Unif(1, 3), λ 2 Unif(1, 3), λ 3 = 0 Tudose & Jucan (2007); Forrester et al. (2008) 27

28 Computer experiments example x x 2 x x x 2 x Bayesian Ψ-optimal design and maximin Latin hypercube design for prior 1 (left) and prior 2 (right). 28

29 Computer experiments example For prior 1: Bayesian optimal design has similar space-filling properties to the LHD Average inter-point distance of 1.43 vs 1.40 Bayesian design has 30% smaller average posterior predictive variance than that of the LHD For prior 2: Design points in the x 3 dimension collapse onto the extremes Bayesian design has 18% smaller average posterior predictive variance than that of the LHD Two-dimensional example 29

30 Ride optimisation: design and response variables Each variant can be described by different mass properties, and has different ride characteristics Mass properties (design variables) units Mass kg Mass distribution % Fr Centre of gravity mm Roll inertia kgm 2 Pitch inertia (damping) kgm 2 Damping can be used to tune the ride performance of a given variant 30

31 10 Variants for study Mass Mass dist. Centre of gravity Roll inertia Running physical tests to assess the performance of each of these variants is time consuming and expensive. Instead, a computer model is run that simulates ride performance 31

32 Model-based design Mass Mass dist CG Roll Damp Tailor the design for Gaussian process prediction for the given variants 32

33 Spatial examples Example 1: Known noise to signal ratio τ 2 = σ 2 ε/σ 2 g Linear mean function f T (x)β = β 0 + β 1x 1 + β 2x 2 Correlation function k λ (x, x ) = exp( λ x x 2), with x i, x j [ 1, 1] 2 Prior distributions β σ 2 g N(0, σ 2 gi), σ 2 g Inverse Gamma(3, 1) and λ Uniform(0.1, 1) τ 2 has fixed values 0,0.5,1,2.5 Optimal designs with n = 10 points and average correlation contours 33

34 τ 2 = 0 τ 2 = x x x x 1 τ 2 = 1 τ 2 = x x x x

35 Spatial examples Example 2: Unknown noise to signal ratio τ 2 Linear mean function f T (x)β = β 0 + β 1x 1 + β 2x 2 Matern correlation function k λ (x, x ) = {2 λ 2 1 Γ(λ 2)} 1 (2λ 1/2 2 d/λ1)λ 2 K λ2 (2λ 1/2 2 x x 2/λ 1) where Γ is the gamma function and K λ2 is the modified Bessel function of the second kind Prior distributions β σ 2 g N(0, σ 2 gi), σ 2 g Inverse Gamma(3, 1) and λ 1 Uniform(0.1, 1) τ 2 Uniform(0, 1) Optimal designs with n = 10 points for λ 2 = 0.5 (exponential correlation) and λ 2 =

36 λ 2 = 0.5 λ 2 = x x x x

37 τ 2 = 0 τ 2 = Density Density correlation correlation τ 2 = 0.5 τ 2 = Density Density correlation correlation 37

38 τ 2 = 0 τ 2 = Density Density correlation correlation τ 2 = 0.5 τ 2 = Density Density correlation correlation 38

39 Key points New designs compromise between minimax (coverage) and maximin (spread) designs Different prior information on signal and noise can lead to different designs (space-filling or linear model) Spread of the design points depends crucially on the distribution of the correlations problems with less uniform correlations across the spatial region produce designs with better space-filling properties 39

40 Sensor placement (1) (a) prediction candidate optimal Linear mean; low correlation 40

41 Sensor placement (2) prediction candidate optimal (b) prediction candidate optimal Linear mean; medium correlation 41

42 Sensor placement (3) (c) prediction candidate optimal Linear mean; high correlation 42

43 Estimation 43

44 Bayesian optimal design for estimation For estimating the trend parameters β, we apply the square error loss function L( β, β; ξ) = β β 2 Best estimate of β is ˆβ = E(β y) (Bayes decision) The objective function to be minimised is Ψ(ξ) = E [{β E(β y)} T {β E(β y)}] π(y)dy Y = E [tr [{β E(β y)} {β E(β y)} T ]] π(y)dy Y = tr [var(β y)] π(y)dy. Y 44

45 An approximation Decompose the objective function Ψ(ξ) = E φ y [var(β φ, y; ξ)] π(y; ξ) dy + var φ y [E(β φ, y; ξ)] π(y; ξ) dy = Ψ 1(ξ) + Ψ 2(ξ) Set Ψ 2(ξ) = 0, and use the approximation Ψ(ξ) Ψ 1(ξ) = E φ y [var β φ, y; ξ)] π(y; ξ) dy = var(β φ, y; ξ)π(y φ; ξ)π(φ) dy dφ tr[f T L 1 n F + R] 1 π(φ) dφ 45

46 Example We assume the model where: s corresponds to treatment factors x corresponds to nuisance factors y(s, x) = m(s) + Z(x) + ɛ m(s) is the mean function and Z(x) a mean zero Gaussian process prior We consider a well-plate with nine fixed locations and in each location a different treatment is applied. See also work on design for nearest neighbour and autocorrelation structures Kiefer and Wynn (1981,1984) 46

47 Example Make a choice of treatments and where to apply them Four treatment factors; linear mean function f T (s)β = β k=1 β k s k Treatments have three levels 1, 0, 1 Taxicab distance between the location k λ (x, x ) = exp( λ x x 1) for x, x [0, 1] 2 Prior for β, σg 2 is Normal Inverse Gamma Priors for λ, τ 2 are unifrom 47

48 Example s 1 s 2 s 3 s

49 Example Frequency Value of the objective function 49

50 Estimation Identifiability and consistency problems when estimating trend parameters in Gaussian Process model Identifiability issues with mean parameters as they are not orthogonal to Gaussian process Kennedy and O Hagan (2001) The identifiability problem can be avoided by making the mean function orthogonal to the Gaussian process prior Achieved by defining a new correlation function that incorporates the orthogonality condition Modify the correlation function and apply our methodology Plumlee and Joseph (2015) 50

51 Summary New model-based designs Tailored for prediction but take account of uncertainty in parameters that require estimation Strongly influenced by the range and the degree of correlation and the choice of mean function Largely insensitive to the choice of prior hyperparameters for σ 2 g Suitable for computer experiments, spatial/spatio-temporal studies, machine learning applications,... Methodology can be used for robust product designs 51

52 References I Adamou, M. (2014). PhD thesis, University of Southampton. Atkinson, A., Donev, A. & Tobias, R. (2007). Optimum Experimental Designs with SAS. Oxford University Press, 2nd edn. Beck, J. & Guillas. S. (2016). JUQ, 4 (1), Berger, J. & De Oliveria, V. & Sanso, B. (2001). JASA, 96, Boukouvalas, A., Cornford, D. & Stehlik, M. (2014). CSDA, 71, Chaloner, K. & Verdinelli, I. (1995). Statistical Science, 10, Cressie, N.A.C. (1993). Statistics for Spatial Data. Wiley. Diggle, P. & Lophaven, S. (2006). Scand. J. Statist., 33, Forrester, A., Sobester, A. & Keane, A. (2008). Engineering Design via Surrogate Modelling. Wiley. Gorodetsky, A. & Marzouk, Y. (2016). JUQ, 4(1), Gramacy, R.B. & Lee, H.K.H. (2009). Technometrics, 51, Gramacy, R.B. & Polson, N.G. (2011). JCGS, 20, Johnson, M.E., Moore, L.M. & Ylvisaker, D. (1990). JSPI, 26,

53 References II Jones, D.R., Schonlau, M. & Welch, W.J. (1998). J. Glob. Opt., 13, Kiefer, J. and Wynn, H.P. (1981). Annals of Statistics, 9, Kiefer, J. and Wynn, H.P. (1984). Annals of Statistics, 12, Leatherman, E.R., Dean, A.M., & Santner, T.J. (2017), Computational Statistics & Data Analysis, 113, Leatherman, E.R., Santner, T.J. & Dean, A.M. (In Press), Statistics and Computing. McKay, M.D., Beckman, R.J. & Conover, W.J. (1979). Technometrics, 21, Meyer, R. and Nachtsheim, C. (1995). Technometrics, 37, Müller, P. & Parmigiani, G. (1995). JASA, 90, O Hagan, A. (1978). JRSSB, 40, Rasmussen, C.E. & Williams, C.K.I. (2006). Gaussian Processes for Machine Learning, MIT Press. Sacks, J., Welch, W.J., Mitchell, T.J. & Wynn, H.P. (1989). Statistical Science, 4,

54 References III Santner, T.J., Williams, B.J. & Notz, W.I. (2003). The Design and Analysis of Computer Experiments. Springer. Sebastiani, P. & Wynn, H.P. (2000). JRSSB, 62, Tudose, L. & Jucan, D. (2007). Ann. Oradea Uni., Wikle, C.K. & Royle, J.A. (2005). Environmetrics, 16, Wu, R. & Kaufman, C.G. (2013). Submitted. Xia, G., Miranda, M.L. & Gelfand, A.E. (2006). Environmetrics, 17, Zimmerman, D.L. (2006). Environmetrics, 17, Zhou, Z. & Stein, M.L. (2005). JSPI, 134, Zhou, Z. & Stein, M.L. (2006). JABES, 11,

55 55

56 Generalised kriging variance S 2 (x n+ ; φ, ξ) S 2 (x n+; φ, ξ) = L n+ L T n,n+l 1 n L n,n+ + (F n+ L T n,n+l 1 n F )(F T L 1 n F + R) 1 (F n+ L T n,n+l 1 n F ) T Derived from conjugate priors on β and σ 2 g R 1 is the prior precision matrix for β F model matrix; F n+ model matrix for prediction points L n = [k λ (x i, x j) + τ 2 δ ij] n i,j=1, L n+ = [k λ (x i,n+, x j,n+) + τ 2 δ ij] n i,j=1, L n,n+ = [k λ (x i, x j,n+) + τ 2 δ ij] i=n,j=n+ i=1,j=1 56

57 Comparison with IMSPE Leatherman et al. (2017) applied a pseudo-bayesian IMSPE objective function where: Φ(ξ) = σ 2 S 2 (x n+; φ, ξ)π(φ)dφ dx n+ S 2 = L n+ L T n,n+l 1 n L n,n+ + (F n+ L T n,n+l 1 n F )(F T L 1 n F ) 1 (F n+ L T n,n+l 1 n F ) T L n+, L n,n+ and K correlation matrices, F n+ and F model matrices Derived from frequentist perspective Equivalent to non-informative prior for β Our approximation Ψ 1(ξ) demonstrates that Φ(ξ) is also a good approximation to the decision theoretic Bayesian approach 57

58 Coordinate exchange algorithm Overview of algorithm: 1. Choose a random starting design 2. For each point, perform a one dimensional optimisation for each coordinate in turn, with all the other coordinates (in all points) remaining fixed 3. The algorithm selects the coordinate value that minimises the objective function 4. Repeat 2 3 for each point and coordinate, until no non-negligible decrease is obtained in the objective function after one complete iteration through the design Meyer and Nachtsheim, (1995) 58

Gaussian Processes for Computer Experiments

Gaussian Processes for Computer Experiments Gaussian Processes for Computer Experiments Jeremy Oakley School of Mathematics and Statistics, University of Sheffield www.jeremy-oakley.staff.shef.ac.uk 1 / 43 Computer models Computer model represented

More information

Design of experiments for generalized linear models with random block e ects

Design of experiments for generalized linear models with random block e ects Design of experiments for generalized linear models with random block e ects Tim Waite timothy.waite@manchester.ac.uk School of Mathematics University of Manchester Tim Waite (U. Manchester) Design for

More information

An introduction to Bayesian statistics and model calibration and a host of related topics

An introduction to Bayesian statistics and model calibration and a host of related topics An introduction to Bayesian statistics and model calibration and a host of related topics Derek Bingham Statistics and Actuarial Science Simon Fraser University Cast of thousands have participated in the

More information

arxiv: v1 [stat.me] 24 May 2010

arxiv: v1 [stat.me] 24 May 2010 The role of the nugget term in the Gaussian process method Andrey Pepelyshev arxiv:1005.4385v1 [stat.me] 24 May 2010 Abstract The maximum likelihood estimate of the correlation parameter of a Gaussian

More information

Approximately Optimal Experimental Design for Heteroscedastic Gaussian Process Models

Approximately Optimal Experimental Design for Heteroscedastic Gaussian Process Models Neural Computing Research Group Aston University Birmingham B4 7ET United Kingdom Tel: +44 (0)121 333 4631 Fax: +44 (0)121 333 4586 http://www.ncrg.aston.ac.uk/ Approximately Optimal Experimental Design

More information

Introduction to emulators - the what, the when, the why

Introduction to emulators - the what, the when, the why School of Earth and Environment INSTITUTE FOR CLIMATE & ATMOSPHERIC SCIENCE Introduction to emulators - the what, the when, the why Dr Lindsay Lee 1 What is a simulator? A simulator is a computer code

More information

arxiv: v1 [stat.me] 10 Jul 2009

arxiv: v1 [stat.me] 10 Jul 2009 6th St.Petersburg Workshop on Simulation (2009) 1091-1096 Improvement of random LHD for high dimensions arxiv:0907.1823v1 [stat.me] 10 Jul 2009 Andrey Pepelyshev 1 Abstract Designs of experiments for multivariate

More information

Weighted space-filling designs

Weighted space-filling designs Weighted space-filling designs Dave Woods D.Woods@southampton.ac.uk www.southampton.ac.uk/ davew Joint work with Veronica Bowman (Dstl) Supported by the Defence Threat Reduction Agency 1 / 34 Outline Motivation

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

Optimal Designs for Gaussian Process Models via Spectral Decomposition. Ofir Harari

Optimal Designs for Gaussian Process Models via Spectral Decomposition. Ofir Harari Optimal Designs for Gaussian Process Models via Spectral Decomposition Ofir Harari Department of Statistics & Actuarial Sciences, Simon Fraser University September 2014 Dynamic Computer Experiments, 2014

More information

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

PILCO: A Model-Based and Data-Efficient Approach to Policy Search PILCO: A Model-Based and Data-Efficient Approach to Policy Search (M.P. Deisenroth and C.E. Rasmussen) CSC2541 November 4, 2016 PILCO Graphical Model PILCO Probabilistic Inference for Learning COntrol

More information

Monitoring Wafer Geometric Quality using Additive Gaussian Process

Monitoring Wafer Geometric Quality using Additive Gaussian Process Monitoring Wafer Geometric Quality using Additive Gaussian Process Linmiao Zhang 1 Kaibo Wang 2 Nan Chen 1 1 Department of Industrial and Systems Engineering, National University of Singapore 2 Department

More information

Why experimenters should not randomize, and what they should do instead

Why experimenters should not randomize, and what they should do instead Why experimenters should not randomize, and what they should do instead Maximilian Kasy Department of Economics, Harvard University Maximilian Kasy (Harvard) Experimental design 1 / 42 project STAR Introduction

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields

Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields 1 Introduction Jo Eidsvik Department of Mathematical Sciences, NTNU, Norway. (joeid@math.ntnu.no) February

More information

Sensitivity analysis in linear and nonlinear models: A review. Introduction

Sensitivity analysis in linear and nonlinear models: A review. Introduction Sensitivity analysis in linear and nonlinear models: A review Caren Marzban Applied Physics Lab. and Department of Statistics Univ. of Washington, Seattle, WA, USA 98195 Consider: Introduction Question:

More information

Kriging by Example: Regression of oceanographic data. Paris Perdikaris. Brown University, Division of Applied Mathematics

Kriging by Example: Regression of oceanographic data. Paris Perdikaris. Brown University, Division of Applied Mathematics Kriging by Example: Regression of oceanographic data Paris Perdikaris Brown University, Division of Applied Mathematics! January, 0 Sea Grant College Program Massachusetts Institute of Technology Cambridge,

More information

STA414/2104 Statistical Methods for Machine Learning II

STA414/2104 Statistical Methods for Machine Learning II STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements

More information

DESIGNING COMPUTER EXPERIMENTS TO DETERMINE ROBUST CONTROL VARIABLES

DESIGNING COMPUTER EXPERIMENTS TO DETERMINE ROBUST CONTROL VARIABLES Statistica Sinica 14(2004), 571-590 DESIGNING COMPUTER EXPERIMENTS TO DETERMINE ROBUST CONTROL VARIABLES Jeffrey S. Lehman 1,ThomasJ.Santner 2 and William I. Notz 2 1 Battelle Memorial Institute and 2

More information

Limit Kriging. Abstract

Limit Kriging. Abstract Limit Kriging V. Roshan Joseph School of Industrial and Systems Engineering Georgia Institute of Technology Atlanta, GA 30332-0205, USA roshan@isye.gatech.edu Abstract A new kriging predictor is proposed.

More information

Introduction to Gaussian Processes

Introduction to Gaussian Processes Introduction to Gaussian Processes Iain Murray murray@cs.toronto.edu CSC255, Introduction to Machine Learning, Fall 28 Dept. Computer Science, University of Toronto The problem Learn scalar function of

More information

Bayesian Prediction of Code Output. ASA Albuquerque Chapter Short Course October 2014

Bayesian Prediction of Code Output. ASA Albuquerque Chapter Short Course October 2014 Bayesian Prediction of Code Output ASA Albuquerque Chapter Short Course October 2014 Abstract This presentation summarizes Bayesian prediction methodology for the Gaussian process (GP) surrogate representation

More information

STA 4273H: Sta-s-cal Machine Learning

STA 4273H: Sta-s-cal Machine Learning STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our

More information

A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging

A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging A Divide-and-Conquer Bayesian Approach to Large-Scale Kriging Cheng Li DSAP, National University of Singapore Joint work with Rajarshi Guhaniyogi (UC Santa Cruz), Terrance D. Savitsky (US Bureau of Labor

More information

Stat 890 Design of computer experiments

Stat 890 Design of computer experiments Stat 890 Design of computer experiments Will introduce design concepts for computer experiments Will look at more elaborate constructions next day Experiment design In computer experiments, as in many

More information

Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine

Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine Mike Tipping Gaussian prior Marginal prior: single α Independent α Cambridge, UK Lecture 3: Overview

More information

The Bayesian approach to inverse problems

The Bayesian approach to inverse problems The Bayesian approach to inverse problems Youssef Marzouk Department of Aeronautics and Astronautics Center for Computational Engineering Massachusetts Institute of Technology ymarz@mit.edu, http://uqgroup.mit.edu

More information

Practical Bayesian Optimization of Machine Learning. Learning Algorithms

Practical Bayesian Optimization of Machine Learning. Learning Algorithms Practical Bayesian Optimization of Machine Learning Algorithms CS 294 University of California, Berkeley Tuesday, April 20, 2016 Motivation Machine Learning Algorithms (MLA s) have hyperparameters that

More information

Global Optimisation with Gaussian Processes. Michael A. Osborne Machine Learning Research Group Department o Engineering Science University o Oxford

Global Optimisation with Gaussian Processes. Michael A. Osborne Machine Learning Research Group Department o Engineering Science University o Oxford Global Optimisation with Gaussian Processes Michael A. Osborne Machine Learning Research Group Department o Engineering Science University o Oxford Global optimisation considers objective functions that

More information

Outline Introduction OLS Design of experiments Regression. Metamodeling. ME598/494 Lecture. Max Yi Ren

Outline Introduction OLS Design of experiments Regression. Metamodeling. ME598/494 Lecture. Max Yi Ren 1 / 34 Metamodeling ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University March 1, 2015 2 / 34 1. preliminaries 1.1 motivation 1.2 ordinary least square 1.3 information

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Use of Design Sensitivity Information in Response Surface and Kriging Metamodels

Use of Design Sensitivity Information in Response Surface and Kriging Metamodels Optimization and Engineering, 2, 469 484, 2001 c 2002 Kluwer Academic Publishers. Manufactured in The Netherlands. Use of Design Sensitivity Information in Response Surface and Kriging Metamodels J. J.

More information

STA414/2104. Lecture 11: Gaussian Processes. Department of Statistics

STA414/2104. Lecture 11: Gaussian Processes. Department of Statistics STA414/2104 Lecture 11: Gaussian Processes Department of Statistics www.utstat.utoronto.ca Delivered by Mark Ebden with thanks to Russ Salakhutdinov Outline Gaussian Processes Exam review Course evaluations

More information

Sequential adaptive designs in computer experiments for response surface model fit

Sequential adaptive designs in computer experiments for response surface model fit Statistics and Applications Volume 6, Nos. &, 8 (New Series), pp.7-33 Sequential adaptive designs in computer experiments for response surface model fit Chen Quin Lam and William I. Notz Department of

More information

Comparing Non-informative Priors for Estimation and Prediction in Spatial Models

Comparing Non-informative Priors for Estimation and Prediction in Spatial Models Environmentrics 00, 1 12 DOI: 10.1002/env.XXXX Comparing Non-informative Priors for Estimation and Prediction in Spatial Models Regina Wu a and Cari G. Kaufman a Summary: Fitting a Bayesian model to spatial

More information

Machine Learning. Bayesian Regression & Classification. Marc Toussaint U Stuttgart

Machine Learning. Bayesian Regression & Classification. Marc Toussaint U Stuttgart Machine Learning Bayesian Regression & Classification learning as inference, Bayesian Kernel Ridge regression & Gaussian Processes, Bayesian Kernel Logistic Regression & GP classification, Bayesian Neural

More information

Bayesian Dynamic Linear Modelling for. Complex Computer Models

Bayesian Dynamic Linear Modelling for. Complex Computer Models Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

Gaussian Process Regression and Emulation

Gaussian Process Regression and Emulation Gaussian Process Regression and Emulation STAT8810, Fall 2017 M.T. Pratola September 22, 2017 Today Experimental Design; Sensitivity Analysis Designing Your Experiment If you will run a simulator model,

More information

Bayesian Geostatistical Design

Bayesian Geostatistical Design Johns Hopkins University, Dept. of Biostatistics Working Papers 6-5-24 Bayesian Geostatistical Design Peter J. Diggle Medical Statistics Unit, Lancaster University, UK & Department of Biostatistics, Johns

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Bayesian Calibration of Inexact Computer Models

Bayesian Calibration of Inexact Computer Models Bayesian Calibration of Inexact Computer Models James Matuk Research Group in Design of Physical and Computer Experiments March 5, 2018 James Matuk (STAT 8750.02) Calibration of Inexact Computer Models

More information

On Gaussian Process Models for High-Dimensional Geostatistical Datasets

On Gaussian Process Models for High-Dimensional Geostatistical Datasets On Gaussian Process Models for High-Dimensional Geostatistical Datasets Sudipto Banerjee Joint work with Abhirup Datta, Andrew O. Finley and Alan E. Gelfand University of California, Los Angeles, USA May

More information

Gaussian Processes 1. Schedule

Gaussian Processes 1. Schedule 1 Schedule 17 Jan: Gaussian processes (Jo Eidsvik) 24 Jan: Hands-on project on Gaussian processes (Team effort, work in groups) 31 Jan: Latent Gaussian models and INLA (Jo Eidsvik) 7 Feb: Hands-on project

More information

A short introduction to INLA and R-INLA

A short introduction to INLA and R-INLA A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate

More information

Graphical Models for Collaborative Filtering

Graphical Models for Collaborative Filtering Graphical Models for Collaborative Filtering Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Sequence modeling HMM, Kalman Filter, etc.: Similarity: the same graphical model topology,

More information

A new Hierarchical Bayes approach to ensemble-variational data assimilation

A new Hierarchical Bayes approach to ensemble-variational data assimilation A new Hierarchical Bayes approach to ensemble-variational data assimilation Michael Tsyrulnikov and Alexander Rakitko HydroMetCenter of Russia College Park, 20 Oct 2014 Michael Tsyrulnikov and Alexander

More information

Gaussian Processes in Machine Learning

Gaussian Processes in Machine Learning Gaussian Processes in Machine Learning November 17, 2011 CharmGil Hong Agenda Motivation GP : How does it make sense? Prior : Defining a GP More about Mean and Covariance Functions Posterior : Conditioning

More information

Variational Methods in Bayesian Deconvolution

Variational Methods in Bayesian Deconvolution PHYSTAT, SLAC, Stanford, California, September 8-, Variational Methods in Bayesian Deconvolution K. Zarb Adami Cavendish Laboratory, University of Cambridge, UK This paper gives an introduction to the

More information

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.)

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.) Prof. Daniel Cremers 2. Regression (cont.) Regression with MLE (Rep.) Assume that y is affected by Gaussian noise : t = f(x, w)+ where Thus, we have p(t x, w, )=N (t; f(x, w), 2 ) 2 Maximum A-Posteriori

More information

Nonparametric Bayesian Methods - Lecture I

Nonparametric Bayesian Methods - Lecture I Nonparametric Bayesian Methods - Lecture I Harry van Zanten Korteweg-de Vries Institute for Mathematics CRiSM Masterclass, April 4-6, 2016 Overview of the lectures I Intro to nonparametric Bayesian statistics

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Fast Bayesian Inference for Computer Simulation Inverse Problems

Fast Bayesian Inference for Computer Simulation Inverse Problems Fast Bayesian Inference for Computer Simulation Inverse Problems Matthew Taddy, Herbert K. H. Lee & Bruno Sansó University of California, Santa Cruz, Department of Applied Mathematics and Statistics January

More information

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling François Bachoc former PhD advisor: Josselin Garnier former CEA advisor: Jean-Marc Martinez Department

More information

Overview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation

Overview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation Overview Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation Probabilistic Interpretation: Linear Regression Assume output y is generated

More information

Bayes: All uncertainty is described using probability.

Bayes: All uncertainty is described using probability. Bayes: All uncertainty is described using probability. Let w be the data and θ be any unknown quantities. Likelihood. The probability model π(w θ) has θ fixed and w varying. The likelihood L(θ; w) is π(w

More information

Gaussian Process Regression

Gaussian Process Regression Gaussian Process Regression 4F1 Pattern Recognition, 21 Carl Edward Rasmussen Department of Engineering, University of Cambridge November 11th - 16th, 21 Rasmussen (Engineering, Cambridge) Gaussian Process

More information

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite

More information

Introduction to Gaussian Processes

Introduction to Gaussian Processes Introduction to Gaussian Processes 1 Objectives to express prior knowledge/beliefs about model outputs using Gaussian process (GP) to sample functions from the probability measure defined by GP to build

More information

Neutron inverse kinetics via Gaussian Processes

Neutron inverse kinetics via Gaussian Processes Neutron inverse kinetics via Gaussian Processes P. Picca Politecnico di Torino, Torino, Italy R. Furfaro University of Arizona, Tucson, Arizona Outline Introduction Review of inverse kinetics techniques

More information

Dynamic System Identification using HDMR-Bayesian Technique

Dynamic System Identification using HDMR-Bayesian Technique Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in

More information

Bayesian methods in economics and finance

Bayesian methods in economics and finance 1/26 Bayesian methods in economics and finance Linear regression: Bayesian model selection and sparsity priors Linear Regression 2/26 Linear regression Model for relationship between (several) independent

More information

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the

More information

On Bayesian Computation

On Bayesian Computation On Bayesian Computation Michael I. Jordan with Elaine Angelino, Maxim Rabinovich, Martin Wainwright and Yun Yang Previous Work: Information Constraints on Inference Minimize the minimax risk under constraints

More information

Gaussian with mean ( µ ) and standard deviation ( σ)

Gaussian with mean ( µ ) and standard deviation ( σ) Slide from Pieter Abbeel Gaussian with mean ( µ ) and standard deviation ( σ) 10/6/16 CSE-571: Robotics X ~ N( µ, σ ) Y ~ N( aµ + b, a σ ) Y = ax + b + + + + 1 1 1 1 1 1 1 1 1 1, ~ ) ( ) ( ), ( ~ ), (

More information

Probabilistic Graphical Models Lecture 20: Gaussian Processes

Probabilistic Graphical Models Lecture 20: Gaussian Processes Probabilistic Graphical Models Lecture 20: Gaussian Processes Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 30, 2015 1 / 53 What is Machine Learning? Machine learning algorithms

More information

Geostatistical Modeling for Large Data Sets: Low-rank methods

Geostatistical Modeling for Large Data Sets: Low-rank methods Geostatistical Modeling for Large Data Sets: Low-rank methods Whitney Huang, Kelly-Ann Dixon Hamil, and Zizhuang Wu Department of Statistics Purdue University February 22, 2016 Outline Motivation Low-rank

More information

Kullback-Leibler Designs

Kullback-Leibler Designs Kullback-Leibler Designs Astrid JOURDAN Jessica FRANCO Contents Contents Introduction Kullback-Leibler divergence Estimation by a Monte-Carlo method Design comparison Conclusion 2 Introduction Computer

More information

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model (& discussion on the GPLVM tech. report by Prof. N. Lawrence, 06) Andreas Damianou Department of Neuro- and Computer Science,

More information

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Abhirup Datta 1 Sudipto Banerjee 1 Andrew O. Finley 2 Alan E. Gelfand 3 1 University of Minnesota, Minneapolis,

More information

Introduction to Gaussian-process based Kriging models for metamodeling and validation of computer codes

Introduction to Gaussian-process based Kriging models for metamodeling and validation of computer codes Introduction to Gaussian-process based Kriging models for metamodeling and validation of computer codes François Bachoc Department of Statistics and Operations Research, University of Vienna (Former PhD

More information

Learning Gaussian Process Models from Uncertain Data

Learning Gaussian Process Models from Uncertain Data Learning Gaussian Process Models from Uncertain Data Patrick Dallaire, Camille Besse, and Brahim Chaib-draa DAMAS Laboratory, Computer Science & Software Engineering Department, Laval University, Canada

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Lecture : Probabilistic Machine Learning

Lecture : Probabilistic Machine Learning Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning

More information

Confidence Estimation Methods for Neural Networks: A Practical Comparison

Confidence Estimation Methods for Neural Networks: A Practical Comparison , 6-8 000, Confidence Estimation Methods for : A Practical Comparison G. Papadopoulos, P.J. Edwards, A.F. Murray Department of Electronics and Electrical Engineering, University of Edinburgh Abstract.

More information

Mustafa H. Tongarlak Bruce E. Ankenman Barry L. Nelson

Mustafa H. Tongarlak Bruce E. Ankenman Barry L. Nelson Proceedings of the 0 Winter Simulation Conference S. Jain, R. R. Creasey, J. Himmelspach, K. P. White, and M. Fu, eds. RELATIVE ERROR STOCHASTIC KRIGING Mustafa H. Tongarlak Bruce E. Ankenman Barry L.

More information

Part 1: Expectation Propagation

Part 1: Expectation Propagation Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud

More information

arxiv: v3 [stat.co] 26 Dec 2018

arxiv: v3 [stat.co] 26 Dec 2018 acebayes: An R Package for Bayesian Optimal Design of Experiments via Approximate Coordinate Exchange Antony M. Overstall University of Southampton David C. Woods University of Southampton Maria Adamou

More information

A Comparison Of Design And. Model Selection Methods. For Supersaturated Experiments

A Comparison Of Design And. Model Selection Methods. For Supersaturated Experiments Working Paper M09/20 Methodology A Comparison Of Design And Model Selection Methods For Supersaturated Experiments Christopher J. Marley, David C. Woods Abstract Various design and model selection methods

More information

Gaussian process for nonstationary time series prediction

Gaussian process for nonstationary time series prediction Computational Statistics & Data Analysis 47 (2004) 705 712 www.elsevier.com/locate/csda Gaussian process for nonstationary time series prediction Soane Brahim-Belhouari, Amine Bermak EEE Department, Hong

More information

acebayes: An R Package for Bayesian Optimal Design of Experiments via Approximate Coordinate Exchange

acebayes: An R Package for Bayesian Optimal Design of Experiments via Approximate Coordinate Exchange acebayes: An R Package for Bayesian Optimal Design of Experiments via Approximate Coordinate Exchange Antony M. Overstall University of Southampton David C. Woods University of Southampton Maria Adamou

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

Bayesian estimation of the discrepancy with misspecified parametric models

Bayesian estimation of the discrepancy with misspecified parametric models Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012

More information

Nearest Neighbor Gaussian Processes for Large Spatial Data

Nearest Neighbor Gaussian Processes for Large Spatial Data Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns

More information

A Bayesian Approach to Prediction and Variable Selection Using Nonstationary Gaussian Processes

A Bayesian Approach to Prediction and Variable Selection Using Nonstationary Gaussian Processes A Bayesian Approach to Prediction and Variable Selection Using Nonstationary Gaussian Processes Dissertation Presented in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy in

More information

The Variational Gaussian Approximation Revisited

The Variational Gaussian Approximation Revisited The Variational Gaussian Approximation Revisited Manfred Opper Cédric Archambeau March 16, 2009 Abstract The variational approximation of posterior distributions by multivariate Gaussians has been much

More information

Carl N. Morris. University of Texas

Carl N. Morris. University of Texas EMPIRICAL BAYES: A FREQUENCY-BAYES COMPROMISE Carl N. Morris University of Texas Empirical Bayes research has expanded significantly since the ground-breaking paper (1956) of Herbert Robbins, and its province

More information

The geometry of Gaussian processes and Bayesian optimization. Contal CMLA, ENS Cachan

The geometry of Gaussian processes and Bayesian optimization. Contal CMLA, ENS Cachan The geometry of Gaussian processes and Bayesian optimization. Contal CMLA, ENS Cachan Background: Global Optimization and Gaussian Processes The Geometry of Gaussian Processes and the Chaining Trick Algorithm

More information

Kriging and Alternatives in Computer Experiments

Kriging and Alternatives in Computer Experiments Kriging and Alternatives in Computer Experiments C. F. Jeff Wu ISyE, Georgia Institute of Technology Use kriging to build meta models in computer experiments, a brief review Numerical problems with kriging

More information

Follow-Up Experimental Designs for Computer Models and Physical Processes

Follow-Up Experimental Designs for Computer Models and Physical Processes This article was downloaded by: [Acadia University] On: 13 June 2012, At: 07:19 Publisher: Taylor & Francis Informa Ltd Registered in England and Wales Registered Number: 1072954 Registered office: Mortimer

More information

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,

More information

Multiple-step Time Series Forecasting with Sparse Gaussian Processes

Multiple-step Time Series Forecasting with Sparse Gaussian Processes Multiple-step Time Series Forecasting with Sparse Gaussian Processes Perry Groot ab Peter Lucas a Paul van den Bosch b a Radboud University, Model-Based Systems Development, Heyendaalseweg 135, 6525 AJ

More information

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Bayesian Learning. Tobias Scheffer, Niels Landwehr

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Bayesian Learning. Tobias Scheffer, Niels Landwehr Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning Tobias Scheffer, Niels Landwehr Remember: Normal Distribution Distribution over x. Density function with parameters

More information

Stochastic Spectral Approaches to Bayesian Inference

Stochastic Spectral Approaches to Bayesian Inference Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to

More information

Statistica Sinica Preprint No: SS R2

Statistica Sinica Preprint No: SS R2 Statistica Sinica Preprint No: SS-2016-0423.R2 Title Construction of Maximin Distance Designs via Level Permutation and Expansion Manuscript ID SS-2016-0423.R2 URL http://www.stat.sinica.edu.tw/statistica/

More information

Gaussian process regression for Sensitivity analysis

Gaussian process regression for Sensitivity analysis Gaussian process regression for Sensitivity analysis GPSS Workshop on UQ, Sheffield, September 2016 Nicolas Durrande, Mines St-Étienne, durrande@emse.fr GPSS workshop on UQ GPs for sensitivity analysis

More information

Lecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu

Lecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu Lecture: Gaussian Process Regression STAT 6474 Instructor: Hongxiao Zhu Motivation Reference: Marc Deisenroth s tutorial on Robot Learning. 2 Fast Learning for Autonomous Robots with Gaussian Processes

More information

Hypervolume-based Multi-objective Bayesian Optimization with Student-t Processes

Hypervolume-based Multi-objective Bayesian Optimization with Student-t Processes Hypervolume-based Multi-objective Bayesian Optimization with Student-t Processes Joachim van der Herten Ivo Coucuyt Tom Dhaene IDLab Ghent University - imec igent Tower - Department of Electronics and

More information