AEROSOL MODEL SELECTION AND UNCERTAINTY MODELLING BY RJMCMC TECHNIQUE

Size: px
Start display at page:

Download "AEROSOL MODEL SELECTION AND UNCERTAINTY MODELLING BY RJMCMC TECHNIQUE"

Transcription

1 AEROSOL MODEL SELECTION AND UNCERTAINTY MODELLING BY RJMCMC TECHNIQUE Marko Laine 1, Johanna Tamminen 1, Erkki Kyrölä 1, and Heikki Haario 2 1 Finnish Meteorological Institute, Helsinki, Finland 2 Lappeenranta University of Technology, Lappeenranta, Finland ABSTRACT We apply Bayesian model selection techniques on GO- MOS inverse problem to obtain information on which type of aerosol model fits best the data and to show how the uncertainty of the aerosol model can be included in the error estimates. We use reversible jump Markov chain Monte Carlo (RJMCMC), which is a general extension to the Metropolis-Hastings MCMC algorithm that can explore parameter space of varying/unknown dimensionality. This makes it suitable for the model determination problem. This work is related to Envisat AO NORM. Key words: Envisat, GOMOS, aerosol model, Bayesian model selection, MCMC, RJMCMC. 1. INTRODUCTION Advances in computer resources and algorithms have made it possible to consider increasingly complicated models for data. In geophysical sciences the estimation of unknowns in large models are usually handled using linearizations and approximations that can have effect on the uncertainty estimates of the retrievals. Such things as the effect of the modelling error or the effect of the discretization on the uncertainty of the results require careful analysis and are best studied with methods that do rely on approximations. Bayesian inference provides an unified and natural framework to consider uncertainty in estimated values and the model uncertainty. In many cases, classical approximative estimation methods can be seen as special cases of Bayesian analyses. In Bayesian inference, the uncertainty of the estimated value is the primary target of the investigation and whenever computationally possible, the results of the analysis is the full multi dimensional posterior probability density of the unknowns. Bayesian methods allow the study of many kinds of uncertainties, including uncertainty in the model itself. Prior information can be pooled from different sources and incorporated statistically correctly, and the correlation structure of the unknowns can be fully explored. For applying Bayesian inference to modelling we have the MCMC methods, which are computer intensive method to simulate possible values of the unknowns from its posterior probability distribution. For account on applying Bayesian MCMC methods in geophysical applications see for example [1]. 2. BAYESIAN MODEL SELECTION Choosing the right model is a complicated matter. The problem clearly can not be solved by purely statistical considerations. However in order to the model be valid, it has to be able to predict the observations. Statistical methods are able to tell which of the possible solution offer the best fit given the set of models to consider, the data observed, and the prior information that is available. In many cases the ground truth is unknown. We could have several speculative alternatives on the physical behaviour of the system, eg. depending on some unknown state of nature in the location under consideration. Then it is reasonable to also model the uncertainty in the model, for example by introducing several alternative models and let the data to decide what of them to use. And if no single model stands out, then this uncertainty must be taken into account in results by averaging the predictions over the models according to their posterior weights. We briefly introduce the main concepts of model determination in Bayesian framework and discuss various probability distributions of the unknowns that are concerned. Let x stand for the unknown variables of primary interest and η (k) for extra unknown model parameters in the k:th model. We want to use observed data y to estimate the unknowns x, η (k) and also infer about the unknown model k. Index k is a label for finite or countable set of pre-selected models. In the GOMOS example presented in section 6 symbol x will stand for constituent line densities and η (k) will be the aerosol cross section parameters for selected four models k = 1,...,4. To apply Bayesian inference we need to assign prior probabilities jointly for all the unknowns, p(x,η (k), k). By the rules of conditional probabilities the joint probability can be written as a product of conditional probabil- Proc. Envisat Symposium 27, Montreux, Switzerland April 27 (ESA SP-636, July 27)

2 ities p(x,η (k), k) = p(x η (k), k)p(η (k) k)p(k), (1) which reveals the hierarchical nature of the unknowns. Priors can be given sequentially by first assigning probabilities for the unknown variables p(x η (k), k) and the model parameters p(η (k) k) in each model, and then prior probabilities for different models, p(k). In addition, we must build the likelihood function, p(y x,η (k), k), giving the distribution of the observations y, from the forward model and statistical description of the noise. The joint posterior distribution for the unknowns x, η (k) and k given the data y follows the Bayes formula and is a product of the likelihood and the priors: p(x,η (k), k y) = p(y x,η(k), k)p(x η (k), k)p(η (k) k)p(k). p(y) (2) Again, the hierarchical structure of the model is seen. The Bayes formula can be used to reverse the conditional probabilities. Given the observed data y we want infer about the unknown x given the model (or set of models), about the unknown model parameters η (k) and also about the unknown model k. The unconditional probability of the data p(y) in the denominator of the Bayes formula is a normalizing constant that makes the resulting distribution to be a probability distribution. Formally it has to be calculated by integrating over all the unknowns, which is usually high dimensional task. This makes the explicit computation of the posterior distribution a challenging problem. Let us next consider the problem of selecting the best model k from set of competing models. Different models can be judged according to the evidence they give to the observations using marginal model probabilities: p(y k) = p(y θ (k), k)p(θ (k) k)p(k) dθ (k), (3) where θ (k) = (x,η (k) ) stands for the vector of all the unknowns except the model index k. The posterior model probabilities can be written with aid of Bayes formula as p(k y) = p(y k)p(k). (4) p(y) Model comparisons are typically done using posterior odds: p(k 1 y) p(k 2 y) = p(y k 1) p(k 1 ) p(y k 2 ) p(k 2 ), (5) where the first term in the right p(y k 1 )/p(y k 2 ) is called the Bayes factor, which gives the relative evidence of model k 1 wrt. model k 2 given by the data y. The calculation of model probability p(k y) or even that of the evidence p(y k) poses some challenges especially if the class of models considered is large and if there is no natural hierarchy between the models that could be used. Several method for the calculations have bee proposed either by using approximations that avoid the problems of high dimensional integration, or by using results of MCMC runs on single models. The RJMCMC method presented in this article, allows for a simple method for constructions the Bayes factors and model posterior single MCMC run done simultaneously on all the selected models. 3. MARKOV CHAIN MONTE CARLO MCMC In the most general settings we are interested in the whole posterior distribution of the unknowns. Some times we only need some statistics of the distribution such as the mean and standard deviation. These all lead to high dimensional integration problems that generally has no closed form solutions. The exception being the maximum a posterior (MAP) estimate that can be calculated as an solution of an optimization problem of finding the peak of the posterior distribution. Numerical methods for calculating high dimensional integrals have to resort to Monte Carlo methods for all but very moderate sized problems, say dimension less that 4. However, an estimation problem can be considered fully solved if we have an algorithm that can simulate values from the posterior distribution of the unknowns as the integrals can then be replaced with sample averages which are easy to calculate. But even standard Monte Carlo techniques are in trouble when the dimension is high because of the sparseness of spaces of high dimensions. The solution is to use methods based on high dimensional random walks. The most important of these simulation algorithms is called Metropolis-Hasting algorithm (MH). It belongs to set of methods called Markov chain Monte Carlo (MCMC), meaning that the resulting simulated values form a stochastic process where a new value can depend on the previous one, thus forming kind of a random walk in the multi dimensional space of the unknowns. As the name of the method indicates, the results of the simulation is called chain. The MH chain is constructed using a proposal distribution that suggests new values for the unknown x. The value is accepted according to a simple rule that ensures the right distribution of the resulting chain. Adaptive methodologies for the choosing of the proposal distributions [2, 3, 4] allow for efficient simulations even in high dimensional situations with large number of unknowns. The MH algorithm has several useful generalizations and special cases for different purposes. For the rest of this section we let θ stand for all the unknowns of our model, for example both the unknown state variables and unknown model parameters. The basic idea behind MH is that instead of needing to compute the values of the posterior p(θ y) directly, we only need to compute ratios of the posteriors in two distinct parameter values, p(θ 2 y)/p(θ 1 y). This cancels out the normalising constant p(y) and also the parts of the likelihood

3 function p(y θ) that do not depend on the unknown θ. With a little more detail, the MH algorithm can be described as a method for generating a chain of possible parameter realisations θ,θ 1,... in such a way that it forms a sample from the posterior distribution p(θ y). Starting from an initial guess θ. In each step i of the algorithm the current value is θ i and we propose a new value θ using a proposal distribution q(θ i, ). As the notation suggests this proposal can depend on the current value θ i. Typically the proposal q is multi dimensional Gaussian distribution, centered at the current stage θ i. The new value is accepted using acceptance probability α(θ i,θ ) that is defined as ( α(θ i,θ ) = min 1, p(y θ )p(θ )q(θ,θ i ) p(y θ i )p(θ i )q(θ i,θ ) ), (6) in which case we set θ i+1 = θ. If θ is not accepted, the chain stays at the current value, that is θ i+1 = θ i, and the chain just repeats itself. If the proposal is symmetric, so q(θ 1,θ 2 ) = q(θ 2,θ 1 ), as it is case with the Gaussian density, we see that qs cancel out and a new value θ it is accepted unconditionally if it is better than the previous value, e.g. if p(θ y)/p(θ i y) > 1. If it is not better in the above sense, then θ is accepted as a new value with probability equal to the posterior ratio p(θ y)/p(θ i y). Using the standard Markov chain theory it is easy to show that this algorithm produces a chain of values that will eventually follow the posterior distribution p(θ y). "Eventually" means that we must allow some burn-in time to let the chain reach the limiting distribution. The MH algorithm can be thought as traveling uphill towards the peak of the posterior distribution, but occasionally taking steps downhill. The percentage of time spent at each altitude of the hill corresponds exactly to the probability of that height. After the MCMC run we have at our disposal a chain of values of the parameter vector. The inference about the unknowns is done with statistics calculated from the chain. The mean of the chain is the best point estimate for the unknown, a histogram or kernel density gives visual estimate of the uncertainty in estimates values, etc. If we think of the generated chain as a matrix where the number of rows corresponds to the size of the MCMC sample and the number of columns corresponds to the number of unknowns in the model, then each row is a possible realization of the model and these appear in correct proportions corresponding to the posterior distribution. 4. REVERSIBLE JUMP MCMC Reversible jump MCMC, RJMCMC is a general extension to the standard MH to allow the chain to explore unknowns of varying dimension [5]. This makes it possible to use MCMC methods for Bayesian model determination. Idea behind RJMCMC is that we can form the proposal distribution in such way that it can perform reversible jumps between spaces of different dimensions. This means for example that the random walk of the MH algorithm can simultaneously explore different models for the same data. The RJMCMC method is becoming a standard statistical machinery when doing analyses with complicated models. Its main advantage is its generality, it extend the standard MH to a variety of problems where the number of unknowns is also unknown, or when there are discrete choices in the model. In GOMOS profile retrieval such would be the aerosol cross section model used, or the number of constituents that are to be inverted from one occultation. The next section describes a version of the RJMCMC algorithm that makes it is possible to write computer code that solves the model determination problem for large class of problems typical in geophysical modelling. 5. IMPLEMENTING RJMCMC RJMCMC algorithm as initiated by Green [5] can be presented in theoretical framework that extends the standard MH algorithm to a very general state space of the unknowns. We will not present the general theory here, but refer to [5]. Instead, we show how the method can be implemented in case when we are contemplating several different models for the same data. This is also based on the work of Green [6] and is called automatic RJMCMC. In automatic RJMCMC [5, 6] a MCMC sampler is constructed that automatically jumps between models. Suppose that for each model k, the target posterior distributions can be approximated by mean vectors µ k and covariance matrices C k = Rk T R k, where R k s are the Cholesky decompositions. We can then use Gaussian distributions with mean µ k and covariance matrix C k, N(µ k, C k ), (or some scaled version of it) as the proposals in MH algorithm. Automatic RJMCMC uses the following transformation of the parameters between the models. Let again θ (k) stand for the vector or all the unknowns in model k. In MCMC setting we can consider it as one row in the chain matrix representing possible realization of unknowns corresponding to model k. A scaled and normalized versions of the chain values can be computed as z k = (θ (k) µ k )R 1 k. (7) If model j has same dimension as model i we have a simple transformation from model space i to model space j as θ ( j) = µ j + z i R j. (8) If the dimensions of two models do not match, we either drop some columns or add new dimensions using independent Gaussian random numbers, u N(, I), and

4 arrive to transformations µ i + [z i ] n j 1 R j if n i > n j θ (i) µ = i + z i R j if n i = n [ ] j zi µ i + R j if n i < n j. u Here [z] j 1 means the first j components of vector z. Next we need a proposal distribution for choosing a new model. Let p(i, j) to be the probability to try to jump to a model j when the chain is currently at the the model i. So if the current model is i the next model in chosen with a draw from distribution p(i, ). Assume that model j is drawn. Then the current parameter vector is transformed to the new model according to equation (9). The acceptance probability for the RJMCMC sampler can be written as written as ( ) α(θ (i),θ ( j) ) = max 1, p(y θ( j), j)p(θ ( j), j)p( j, i) R j p(y θ (i), i)p(θ (i), i)p(i, j) R i g, T(λ) = exp (1) where R is the determinant of matrix R and g = φ(u) if n i > n j, g = φ(u) 1 if n i < n j, and g = 1 if n i = n j, φ being the probability distribution function of independent multi dimensional Gaussian values, N(, I). (9) This sampler is easy to implement but its success depends on how good the Gaussian approximations are to provide decent proposals from model to model move. It is however typical in many geophysical applications to have uncertainty that is approximately Gaussian. This is the reason for the classical estimation methods to work so well as they many times do. Using RJMCMC allows us to use Bayesian model selection methods and incorporate prior information, such as positivity constraints in a statistically sound manner. Also we recover the nonlinear correlation structures between the uncertainties that usually are not found by other than with fully Bayesian methods. 6. GOMOS AEROSOL MODEL To demonstrate the use of RJMCMC, we apply the automatic RJMCMC method to aerosol model in the GO- MOS retrieval. The forward model is the standard GO- MOS model for the spectral transmission according to the Beers law. The cross section that is used for aerosol line density is however just an approximations of the underlying aerosol extinctions process that depends on lot of different factors. That is why an approximate analytical approximation of the wavelength dependency is used. It is typically done by using a function that behaves like 1/λ, where λ is the wavelength. See [7] for a comparison of different aerosol extinction models for GOMOS inversion using simulated transmission data. We consider four different aerosol cross section models: the standard 1/λ model (model 1), second degree polynomial on λ (model 2), 1/λ 2 dependence (model 3), and a second degree polynomial on 1/λ (model 4). The aerosol models are parametrized using the aerosol extinction at 5 nm (models 1 and 2) or at 3, 5 and 6 nm (models 3 and 4). A positivity prior is put on these values. We concentrate on inverting the integrated line densities from the transmission spectra. This is called the spectral inversion step in GOMOS literature. The vertical inversion of transforming the line densities to the actual constituent densities is a linear operation that is done after the line densities for all the heights have been inverted and it is not considered here. Let N be the vector of integrated line densities of the constituents to be retrieved (O 3, NO 2, NO 3, air, aerosols) and matrix α the corresponding cross sections. The cross section of aerosol depends on the model parameters η (k). The forward model for the observed transmission T is ( ) α(η (k) )N +ǫ(λ), with ǫ(λ) N(,σ 2 wλ 2 ). (11) As the chosen aerosol model will affect the size of the residuals, the error variance is assumed to be of form σ 2 (λ) = σ 2 w 2 λ with known weights w λ for each wavelength λ and model dependent unknown scalar σ 2, which is also estimated by MCMC. The likelihood function is thus seen to be ( p(t N,η (k) ) exp 1 ) 2σ 2 SS(N,η(k) ), (12) where SS(N,η (k) ) is the weighted sum of squares function ( ( T(λ) exp α(η (k) )N ) SS(N,η (k) ) = λ w λ ) 2. (13) As for priors, only positivity constraint for the line densities is used. As the cross section for air closely resembles that of the aerosols, more informative prior for the density of air could be used to identify the aerosol model more easily. For the unknown error variance factor σ 2 non informative inverse Gamma prior is used. All the four model are taken a priori to be equally likely Results The RJMCMC method is so far tried on a limited number of selected occultations and no definitive analyses have been done. For each line of sight, and with given aerosol model, the problem of inverting the line densities from the transmittance is a nonlinear problem with 5 unknowns. So this is fairly easy problem given appropriate initial guesses and if the noise level is not overwhelming. The estimation problem can be solved in least squares sense as nonlinear optimization problem, using eg. Levenberg-Marquardt method. This is also an easy problem for standard standard MCMC simulation.

5 1 posterior probabiliities of the four aerosol models 3.28 x 12 O3 1 x 116 NO probability x NO x 126 Air Aerosols model probabilities height [km] Figure 1. RJMCMC run is done for each height in one GOMOS occultation. The posterior model probabilities are calculated for the four models at each height. The colours show how the different cross sections models get weight depending on the altitude. The colouring is the same as in figures 2 and 3, Model 1: red, model 2 green, model 3 blue, model 4 magenta Figure 2. MCMC chain from GOMOS line density inversion with different aerosol cross section models. The plot labeled "aerosols" is the relative optical extinction at 5 nm. Last plot shows the posterior model probabilities p(k y). MCMC can even be extended to a one step solution, where all the heights are solved simultaneously, with regularization (smoothness) priors on the vertical structure of the profiles. To use RJMCMC with model selection we use the following strategy. First, for each occultation height and for each aerosol model MCMC runs are done separately using delayed rejection adaptive Metropolis algorithm [4] for robust and efficient way to find the individual posterior distributions. From the MCMC chains of these runs the mean vectors and covariance matrices together with their Cholesky factors are calculated to produce µ i, and R i, i = 1,...,4 that are needed in the RJMCMC stage. Secondly, RJMCMC run is done for a chain of length 5 using automatic RJMCMC algorithm. The resulting chains are visually investigated using 1-D plots like in figure 2 to judge if the chains have converged. For model selection we calculate the relative times the RJMCMC has spent in each model. In figure 1 the results for each altitude for the chosen occultation is shown. In the most of the heights one model clearly stands out as the only one that fits the data using the criteria of model posterior probability. As an illustration on model averaging we choose one altitude at 18 km where all the four models have gained some posterior probability. Figures 2, 3 and 4 show results on this selected altitude. Figure 2 shows the MCMC chains for the line densities for one selected GOMOS occultation. The horizontal axis runs with the simulation indexes, vertical axis being the simulated and accepted values for the line density for each constituent. The color indicates in witch model the algorithm is in each step. Plot on the lower left cor- ner labeled "Aerosol" show the relative aerosol extinction at 5 nm for all models. The last plot shows relative times spent in each model. Of the total 5 MCMC simulations the model 1 is visited 129 times, model , model and model times, which makes the corresponding marginal model posterior probabilities p(k i y), i = 1,...,4 to be.3,.321,.634, and.43. Figure 3 gives the estimated posterior densities describing the uncertainties in the line densities. Marginal posteriors for each three models are shown separately. The posteriors are combined to form an uncertainty estimate that takes into account also the uncertainty in the model. The posterior probabilities of the models are just the relative times the chain has spent on each model. This depend on given prior weights for each model and on how well each different model fit the data compared to other models. In the present example, all the models are taken a priori to be equally likely, so p(k) = 1/4 for k = 1,...,4. Model averaging is useful when we are not able to get the best model. The model is then a mixture of different models each weighted according to their posterior weight. The uncertainty in the model is taken into account in the predictions and in the posterior inference of the constituent values. Lastly, to study the aerosol cross sections estimated by the four models considered here, we calculate the optical extinctions for all the aerosol model parameters in the MCMC chain. Then, for each wavelength, we can check the distribution of values of the extinctions realised in the chain and from these values calculate the posteroir limits of possible values. The resulting predictive envelopes are in figure 4.

6 7. CONCLUSIONS O NO3 x x 1 15 model 1 model 2 model 3 model 4 averaged model NO Air x Figure 3. Posterior density estimates of the constituent line densities calculated from the MCMC chains. The thicker line is the uncertainty coming from the averaged model that takes into account the model uncertainty. x 1 26 RJMCMC provides simulation based method for model determination and uncertainty analysis. If one model clearly stands out, then we can select it as the true model, but if the data does not give any indication on the right model, and no accurate prior for the true model is available, then the uncertainty in the model has to be taken into account in the model predictions. The study presented here is still "a proof of concept" for the RJMCMC method, and no comprehensive runs have been done yet to say any definitive results on the right GOMOS aerosol model. The study of aerosols in GOMOS inversion is further complicated by the fact that, in addition to aerosols, parts of the unmodelled variations in the GOMOS spectra are due to the scintillation effects caused by turbulence. The effects are actively studied in the FMI at the moment, and the current project will hopefully give useful methodological tools for these studies. Acknowledgments This work is done under financial support from Finnish Funding Agency for Technology and Innovation (Tekes) within project MASI Modelling and simulation. REFERENCES model model model model Figure 4. Estimated aerosol extinctions for the selected altitude of the example given in the text. Solid line is the fitted median cross section. Grey areas correspond to 5%, 95% and 95% posterior limits of the extinctions. [1] J. Tamminen and E. Kyrölä. Bayesian solution for nonlinear and non-gaussian inverse problem by Markov chain Monte Carlo method. Journal of Geophysical Research, 16(D13): , 21. [2] H. Haario, E. Saksman, and J. Tamminen. An adaptive Metropolis algorithm. Bernoulli, 7(2): , 21. [3] H. Haario, M. Laine, M. Lehtinen, E. Saksman, and J. Tamminen. MCMC methods for high dimensional inversion in remote sensing. Journal of the Royal Statistical Society, Series B, 66(3):591 67, 24. [4] H. Haario, M. Laine, A. Mira, and E. Saksman. DRAM: Efficient adaptive MCMC. Statistics and Computing, 16: , 26. [5] P. J. Green. Reversible jump Markov chain Monte Carlo computation and Bayesian model determination. Biometrika, 82(4): , [6] P. J. Green. Trans-dimensional Markov chain Monte Carlo. In P. J. Green, N. L. Hjort, and S. Richardson, editors, Highly Structured Stochastic Systems, number 27 in Oxford Statistical Science Series. Oxford University Press, 23. [7] F. Vanhellemont, D. Fussen, J. Dodion, C. Bingen, and N. Mateshvili. Choosing a suitable analytical model for aerosol extinction spectra in the retrieval of UV/visible satellite occultation measurements. Journal of Geophysical Research, 111(D2323), 26.

Markov chain Monte Carlo methods in atmospheric remote sensing

Markov chain Monte Carlo methods in atmospheric remote sensing 1 / 45 Markov chain Monte Carlo methods in atmospheric remote sensing Johanna Tamminen johanna.tamminen@fmi.fi ESA Summer School on Earth System Monitoring and Modeling July 3 Aug 11, 212, Frascati July,

More information

Inverse problems and uncertainty quantification in remote sensing

Inverse problems and uncertainty quantification in remote sensing 1 / 38 Inverse problems and uncertainty quantification in remote sensing Johanna Tamminen Finnish Meterological Institute johanna.tamminen@fmi.fi ESA Earth Observation Summer School on Earth System Monitoring

More information

Advanced uncertainty evaluation of climate models by Monte Carlo methods

Advanced uncertainty evaluation of climate models by Monte Carlo methods Advanced uncertainty evaluation of climate models by Monte Carlo methods Marko Laine marko.laine@fmi.fi Pirkka Ollinaho, Janne Hakkarainen, Johanna Tamminen, Heikki Järvinen (FMI) Antti Solonen, Heikki

More information

DRAGON ADVANCED TRAINING COURSE IN ATMOSPHERE REMOTE SENSING. Inversion basics. Erkki Kyrölä Finnish Meteorological Institute

DRAGON ADVANCED TRAINING COURSE IN ATMOSPHERE REMOTE SENSING. Inversion basics. Erkki Kyrölä Finnish Meteorological Institute Inversion basics y = Kx + ε x ˆ = (K T K) 1 K T y Erkki Kyrölä Finnish Meteorological Institute Day 3 Lecture 1 Retrieval techniques - Erkki Kyrölä 1 Contents 1. Introduction: Measurements, models, inversion

More information

Computer Practical: Metropolis-Hastings-based MCMC

Computer Practical: Metropolis-Hastings-based MCMC Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov

More information

UV/VIS Limb Retrieval

UV/VIS Limb Retrieval UV/VIS Limb Retrieval Erkki Kyrölä Finnish Meteorological Institute 1. Data selection 2. Forward and inverse possibilities 3. Occultation: GOMOS inversion 4. Limb scattering: OSIRIS inversion 5. Summary

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Department of Statistics The University of Auckland https://www.stat.auckland.ac.nz/~brewer/ Emphasis I will try to emphasise the underlying ideas of the methods. I will not be teaching specific software

More information

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

A note on Reversible Jump Markov Chain Monte Carlo

A note on Reversible Jump Markov Chain Monte Carlo A note on Reversible Jump Markov Chain Monte Carlo Hedibert Freitas Lopes Graduate School of Business The University of Chicago 5807 South Woodlawn Avenue Chicago, Illinois 60637 February, 1st 2006 1 Introduction

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Bayesian Phylogenetics:

Bayesian Phylogenetics: Bayesian Phylogenetics: an introduction Marc A. Suchard msuchard@ucla.edu UCLA Who is this man? How sure are you? The one true tree? Methods we ve learned so far try to find a single tree that best describes

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

Assessing Regime Uncertainty Through Reversible Jump McMC

Assessing Regime Uncertainty Through Reversible Jump McMC Assessing Regime Uncertainty Through Reversible Jump McMC August 14, 2008 1 Introduction Background Research Question 2 The RJMcMC Method McMC RJMcMC Algorithm Dependent Proposals Independent Proposals

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

STA 294: Stochastic Processes & Bayesian Nonparametrics

STA 294: Stochastic Processes & Bayesian Nonparametrics MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a

More information

Markov chain Monte Carlo methods for high dimensional inversion in remote sensing

Markov chain Monte Carlo methods for high dimensional inversion in remote sensing J. R. Statist. Soc. B (04) 66, Part 3, pp. 591 7 Markov chain Monte Carlo methods for high dimensional inversion in remote sensing H. Haario and M. Laine, University of Helsinki, Finland M. Lehtinen, University

More information

Randomize-Then-Optimize: A Method for Sampling from Posterior Distributions in Nonlinear Inverse Problems

Randomize-Then-Optimize: A Method for Sampling from Posterior Distributions in Nonlinear Inverse Problems Randomize-Then-Optimize: A Method for Sampling from Posterior Distributions in Nonlinear Inverse Problems The MIT Faculty has made this article openly available. Please share how this access benefits you.

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Supplementary Note on Bayesian analysis

Supplementary Note on Bayesian analysis Supplementary Note on Bayesian analysis Structured variability of muscle activations supports the minimal intervention principle of motor control Francisco J. Valero-Cuevas 1,2,3, Madhusudhan Venkadesan

More information

MCMC for big data. Geir Storvik. BigInsight lunch - May Geir Storvik MCMC for big data BigInsight lunch - May / 17

MCMC for big data. Geir Storvik. BigInsight lunch - May Geir Storvik MCMC for big data BigInsight lunch - May / 17 MCMC for big data Geir Storvik BigInsight lunch - May 2 2018 Geir Storvik MCMC for big data BigInsight lunch - May 2 2018 1 / 17 Outline Why ordinary MCMC is not scalable Different approaches for making

More information

MARKOV CHAIN MONTE CARLO

MARKOV CHAIN MONTE CARLO MARKOV CHAIN MONTE CARLO RYAN WANG Abstract. This paper gives a brief introduction to Markov Chain Monte Carlo methods, which offer a general framework for calculating difficult integrals. We start with

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems

Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems John Bardsley, University of Montana Collaborators: H. Haario, J. Kaipio, M. Laine, Y. Marzouk, A. Seppänen, A. Solonen, Z.

More information

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference

Bayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference 1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE

More information

Reminder of some Markov Chain properties:

Reminder of some Markov Chain properties: Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

LECTURE 15 Markov chain Monte Carlo

LECTURE 15 Markov chain Monte Carlo LECTURE 15 Markov chain Monte Carlo There are many settings when posterior computation is a challenge in that one does not have a closed form expression for the posterior distribution. Markov chain Monte

More information

Learning the hyper-parameters. Luca Martino

Learning the hyper-parameters. Luca Martino Learning the hyper-parameters Luca Martino 2017 2017 1 / 28 Parameters and hyper-parameters 1. All the described methods depend on some choice of hyper-parameters... 2. For instance, do you recall λ (bandwidth

More information

Monte Carlo in Bayesian Statistics

Monte Carlo in Bayesian Statistics Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

1 Using standard errors when comparing estimated values

1 Using standard errors when comparing estimated values MLPR Assignment Part : General comments Below are comments on some recurring issues I came across when marking the second part of the assignment, which I thought it would help to explain in more detail

More information

A Review of Pseudo-Marginal Markov Chain Monte Carlo

A Review of Pseudo-Marginal Markov Chain Monte Carlo A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the

More information

Bayesian Gaussian Process Regression

Bayesian Gaussian Process Regression Bayesian Gaussian Process Regression STAT8810, Fall 2017 M.T. Pratola October 7, 2017 Today Bayesian Gaussian Process Regression Bayesian GP Regression Recall we had observations from our expensive simulator,

More information

Likelihood-free MCMC

Likelihood-free MCMC Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte

More information

F denotes cumulative density. denotes probability density function; (.)

F denotes cumulative density. denotes probability density function; (.) BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models

More information

STA 4273H: Sta-s-cal Machine Learning

STA 4273H: Sta-s-cal Machine Learning STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our

More information

David Giles Bayesian Econometrics

David Giles Bayesian Econometrics David Giles Bayesian Econometrics 1. General Background 2. Constructing Prior Distributions 3. Properties of Bayes Estimators and Tests 4. Bayesian Analysis of the Multiple Regression Model 5. Bayesian

More information

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

Development of Stochastic Artificial Neural Networks for Hydrological Prediction Development of Stochastic Artificial Neural Networks for Hydrological Prediction G. B. Kingston, M. F. Lambert and H. R. Maier Centre for Applied Modelling in Water Engineering, School of Civil and Environmental

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1 Parameter Estimation William H. Jefferys University of Texas at Austin bill@bayesrules.net Parameter Estimation 7/26/05 1 Elements of Inference Inference problems contain two indispensable elements: Data

More information

A Statistical Input Pruning Method for Artificial Neural Networks Used in Environmental Modelling

A Statistical Input Pruning Method for Artificial Neural Networks Used in Environmental Modelling A Statistical Input Pruning Method for Artificial Neural Networks Used in Environmental Modelling G. B. Kingston, H. R. Maier and M. F. Lambert Centre for Applied Modelling in Water Engineering, School

More information

The Metropolis-Hastings Algorithm. June 8, 2012

The Metropolis-Hastings Algorithm. June 8, 2012 The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

MCMC notes by Mark Holder

MCMC notes by Mark Holder MCMC notes by Mark Holder Bayesian inference Ultimately, we want to make probability statements about true values of parameters, given our data. For example P(α 0 < α 1 X). According to Bayes theorem:

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing

More information

Introduction to Probabilistic Machine Learning

Introduction to Probabilistic Machine Learning Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

Lecture 7 and 8: Markov Chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani

More information

Correlation integral likelihood for stochastic differential equations

Correlation integral likelihood for stochastic differential equations Correlation integral likelihood for stochastic differential equations Heikki Haario, Janne Hakkarainen, Ramona Maraia, Sebastian Springer Abstract A new approach was recently introduced for the task of

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

Frequentist-Bayesian Model Comparisons: A Simple Example

Frequentist-Bayesian Model Comparisons: A Simple Example Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal

More information

Introduction. Chapter 1

Introduction. Chapter 1 Chapter 1 Introduction In this book we will be concerned with supervised learning, which is the problem of learning input-output mappings from empirical data (the training dataset). Depending on the characteristics

More information

SAMPLING ALGORITHMS. In general. Inference in Bayesian models

SAMPLING ALGORITHMS. In general. Inference in Bayesian models SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be

More information

Approximate Bayesian Computation

Approximate Bayesian Computation Approximate Bayesian Computation Michael Gutmann https://sites.google.com/site/michaelgutmann University of Helsinki and Aalto University 1st December 2015 Content Two parts: 1. The basics of approximate

More information

Notes on pseudo-marginal methods, variational Bayes and ABC

Notes on pseudo-marginal methods, variational Bayes and ABC Notes on pseudo-marginal methods, variational Bayes and ABC Christian Andersson Naesseth October 3, 2016 The Pseudo-Marginal Framework Assume we are interested in sampling from the posterior distribution

More information

BAYESIAN MODEL CRITICISM

BAYESIAN MODEL CRITICISM Monte via Chib s BAYESIAN MODEL CRITICM Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

Introduction to Bayesian methods in inverse problems

Introduction to Bayesian methods in inverse problems Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction

More information

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007 Bayesian inference Fredrik Ronquist and Peter Beerli October 3, 2007 1 Introduction The last few decades has seen a growing interest in Bayesian inference, an alternative approach to statistical inference.

More information

16 : Markov Chain Monte Carlo (MCMC)

16 : Markov Chain Monte Carlo (MCMC) 10-708: Probabilistic Graphical Models 10-708, Spring 2014 16 : Markov Chain Monte Carlo MCMC Lecturer: Matthew Gormley Scribes: Yining Wang, Renato Negrinho 1 Sampling from low-dimensional distributions

More information

Non-homogeneous Markov Mixture of Periodic Autoregressions for the Analysis of Air Pollution in the Lagoon of Venice

Non-homogeneous Markov Mixture of Periodic Autoregressions for the Analysis of Air Pollution in the Lagoon of Venice Non-homogeneous Markov Mixture of Periodic Autoregressions for the Analysis of Air Pollution in the Lagoon of Venice Roberta Paroli 1, Silvia Pistollato, Maria Rosa, and Luigi Spezia 3 1 Istituto di Statistica

More information

Approximate Bayesian Computation: a simulation based approach to inference

Approximate Bayesian Computation: a simulation based approach to inference Approximate Bayesian Computation: a simulation based approach to inference Richard Wilkinson Simon Tavaré 2 Department of Probability and Statistics University of Sheffield 2 Department of Applied Mathematics

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

Markov chain Monte Carlo

Markov chain Monte Carlo Markov chain Monte Carlo Karl Oskar Ekvall Galin L. Jones University of Minnesota March 12, 2019 Abstract Practically relevant statistical models often give rise to probability distributions that are analytically

More information

Statistical Methods in Particle Physics Lecture 1: Bayesian methods

Statistical Methods in Particle Physics Lecture 1: Bayesian methods Statistical Methods in Particle Physics Lecture 1: Bayesian methods SUSSP65 St Andrews 16 29 August 2009 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan

More information

Bayesian Semiparametric GARCH Models

Bayesian Semiparametric GARCH Models Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

MCMC and Gibbs Sampling. Kayhan Batmanghelich

MCMC and Gibbs Sampling. Kayhan Batmanghelich MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction

More information

Non-Parametric Bayes

Non-Parametric Bayes Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian

More information

Bayesian Semiparametric GARCH Models

Bayesian Semiparametric GARCH Models Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods

More information

Bayesian Estimation with Sparse Grids

Bayesian Estimation with Sparse Grids Bayesian Estimation with Sparse Grids Kenneth L. Judd and Thomas M. Mertens Institute on Computational Economics August 7, 27 / 48 Outline Introduction 2 Sparse grids Construction Integration with sparse

More information

DAG models and Markov Chain Monte Carlo methods a short overview

DAG models and Markov Chain Monte Carlo methods a short overview DAG models and Markov Chain Monte Carlo methods a short overview Søren Højsgaard Institute of Genetics and Biotechnology University of Aarhus August 18, 2008 Printed: August 18, 2008 File: DAGMC-Lecture.tex

More information

Bayesian Machine Learning

Bayesian Machine Learning Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 2: Bayesian Basics https://people.orie.cornell.edu/andrew/orie6741 Cornell University August 25, 2016 1 / 17 Canonical Machine Learning

More information

Lecture : Probabilistic Machine Learning

Lecture : Probabilistic Machine Learning Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning

More information

Choosing among models

Choosing among models Eco 515 Fall 2014 Chris Sims Choosing among models September 18, 2014 c 2014 by Christopher A. Sims. This document is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported

More information

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo (MCMC) Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can

More information

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31 Hierarchical models Dr. Jarad Niemi Iowa State University August 31, 2017 Jarad Niemi (Iowa State) Hierarchical models August 31, 2017 1 / 31 Normal hierarchical model Let Y ig N(θ g, σ 2 ) for i = 1,...,

More information

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 18-16th March Arnaud Doucet

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 18-16th March Arnaud Doucet Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 18-16th March 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Trans-dimensional Markov chain Monte Carlo. Bayesian model for autoregressions.

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Strong Lens Modeling (II): Statistical Methods

Strong Lens Modeling (II): Statistical Methods Strong Lens Modeling (II): Statistical Methods Chuck Keeton Rutgers, the State University of New Jersey Probability theory multiple random variables, a and b joint distribution p(a, b) conditional distribution

More information

The Recycling Gibbs Sampler for Efficient Learning

The Recycling Gibbs Sampler for Efficient Learning The Recycling Gibbs Sampler for Efficient Learning L. Martino, V. Elvira, G. Camps-Valls Universidade de São Paulo, São Carlos (Brazil). Télécom ParisTech, Université Paris-Saclay. (France), Universidad

More information

Using Model Selection and Prior Specification to Improve Regime-switching Asset Simulations

Using Model Selection and Prior Specification to Improve Regime-switching Asset Simulations Using Model Selection and Prior Specification to Improve Regime-switching Asset Simulations Brian M. Hartman, PhD ASA Assistant Professor of Actuarial Science University of Connecticut BYU Statistics Department

More information

Bayesian Inference in Astronomy & Astrophysics A Short Course

Bayesian Inference in Astronomy & Astrophysics A Short Course Bayesian Inference in Astronomy & Astrophysics A Short Course Tom Loredo Dept. of Astronomy, Cornell University p.1/37 Five Lectures Overview of Bayesian Inference From Gaussians to Periodograms Learning

More information

Markov chain Monte Carlo

Markov chain Monte Carlo 1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop

More information

MCMC: Markov Chain Monte Carlo

MCMC: Markov Chain Monte Carlo I529: Machine Learning in Bioinformatics (Spring 2013) MCMC: Markov Chain Monte Carlo Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2013 Contents Review of Markov

More information

Dynamic System Identification using HDMR-Bayesian Technique

Dynamic System Identification using HDMR-Bayesian Technique Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in

More information

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models Matthew S. Johnson New York ASA Chapter Workshop CUNY Graduate Center New York, NY hspace1in December 17, 2009 December

More information

SAMSI Astrostatistics Tutorial. More Markov chain Monte Carlo & Demo of Mathematica software

SAMSI Astrostatistics Tutorial. More Markov chain Monte Carlo & Demo of Mathematica software SAMSI Astrostatistics Tutorial More Markov chain Monte Carlo & Demo of Mathematica software Phil Gregory University of British Columbia 26 Bayesian Logical Data Analysis for the Physical Sciences Contents:

More information