Spatial point processes: Theory and practice illustrated with R

Size: px
Start display at page:

Download "Spatial point processes: Theory and practice illustrated with R"

Transcription

1 Spatial point processes: Theory and practice illustrated with R Department of Mathematical Sciences Aalborg University Lecture IV, February 24, /28

2 Contents of lecture IV Papangelou conditional intensity. Simulating Gibbs models. Maximum pseudolikelihood. Fitting Gibbs models. Residuals and model validation. 2/28

3 Papangelou conditional intensity The main tool for analysing a Gibbs point process is the Papangelou conditional intensity λ(u, X). Informally, the conditional probability of finding a point of the process inside an infinitesimal neighbourhood of the location u, given the complete point pattern at all other locations, is λ(u, X) du. For point processes in a bounded window, the conditional intensity at a location u given the configuration x is related to the probability density f by λ(u, x) = f (x {u}) f (x) (for u x), the ratio of the probability densities for the configuration x with and without the point u added. 3/28

4 Papangelou conditional intensity The Poisson process with intensity function λ(u) has conditional intensity λ(u, x) = λ(u). The conditional intensity for a Poisson process does not depend on the configuration x, because the points of a Poisson process are independent. For the general pairwise interaction process the conditional intensity is n(x) λ(u, x) = b(u) c(u, x i ). i=1 4/28

5 Papangelou conditional intensity For the hard core process, { β if u xi > r for all i λ(u, x) = 0 otherwise which has the nice interpretation that a point u is either permitted or not permitted depending on whether it satisfies the hard core requirement. For the Strauss process λ(u, x) = βγ t(u,x) where t(u, x) = s(x {u}) s(x) is the number of points of x that lie within a distance r of the location u. For γ < 1, this has the interpretation that a random point is less likely to occur at the location u if there are many points in the neighbourhood. (1) 5/28

6 Papangelou conditional intensity For the area-interaction process, λ(u, x) = βγ B(u,x) where B(u, x) = A(x {u}) A(x) is the area of that part of the disc of radius r centred on u that is not covered by discs of radius r centred at the other points x i x. Strauss area interaction + + 6/28

7 Simulating Gibbs models Gibbs models can be simulated by Markov chain Monte Carlo algorithms. (Actually, MCMC algorithms were invented to simulate Gibbs processes.) Currently spatstat offers the function rmh which simulates Gibbs processes using the Metropolis-Hastings algorithm. > rmh(model, start, control) model determines the point process model to be simulated (see help(rmhmodel)). start determines the initial state of the Markov chain (see help(rmhstart)). control specifies control parameters for running the Markov chain, such as the number of iteration steps (see help(rmhcontrol)). 7/28

8 Simulating Gibbs models In the simplest uses of rmh, the three arguments are lists of parameter values. To generate a simulated realisation of the Strauss process with parameters β = 2, γ = 0.7, r = 0.7 in a square of side 10, > mo <- list(cif = "strauss", par = list(beta = 2, + gamma = 0.2, r = 0.7), w = square(10)) > X <- rmh(model = mo, start = list(n.start = 42), + control = list(nrep = 1e+06)) The other arguments specify a random initial state of 42 points, and that the algorithm shall be run for a million iterations. 8/28

9 9/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Simulating Gibbs models Specifically for the Strauss model spatstat offers the possibility of using a perfect simulation algorithm to generate a realisation from the model: > Y <- rstrauss(beta = 2, gamma = 0.2, R = 0.7, + W = square(10)) > plot(x, main = "") > plot(y, main = "")

10 Maximum pseudolikelihood Maximum likelihood estimation is in general very difficult for point process models, and an alternative is to maximise the log pseudolikelihood log PL (θ; x) = log λ(x i ; x) λ(u, x) du. (2) i W In general it is not a likelihood, but the analogue of the score equation log PL (θ) = 0 θ is an unbiased estimating equation. Thus the maximum pseudolikelihood estimator is asymptotically unbiased, consistent and asymptotically normal under appropriate conditions. 10/28

11 Maximum pseudolikelihood Consider any quadrature scheme to approximate the integral m λ(u, x) du λ θ (u j, x)w j W where u j, j = 1,..., m, are points in W and w j > 0 are quadrature weights summing to W. If the quadrature points are chosen such that {x 1,..., x n } {u 1,..., u m } we can write m log PL (θ; x) = (y j log λ j λ j )w j, j=1 j=1 where λ j = λ θ (u j, x) and y j = z j /w j, and { 1 if uj is a data point, u z j = j x 0 if u j is a dummy point, u j / x. This is the log-likelihood of independent Poisson variables Y j with means λ j taken with weights w j. 11/28

12 Fitting Gibbs models in spatstat The function ppm implements maximum pseudolikelihood inference for general Gibbs processes based on the conditional intensity of the model, λ θ (u, x). The model must be loglinear in the parameters θ: log λ θ (u, x) = θ S(u, x), (3) where S(u, x) is a real-valued or vector-valued function of location u and configuration x. Parameters θ appearing in the loglinear form (2) are called regular parameters, and all other parameters are irregular parameters. 12/28

13 Fitting Gibbs models in spatstat For example, the Strauss process conditional intensity can be recast as log λ(u, x) = log β + (log γ)t(u, x) so that θ = (log β, log γ) are regular parameters, but the interaction distance r is an irregular parameter. In spatstat the conditional intensity is split into first-order and higher-order terms: log λ θ (u, x) = η S(u) + ϕ V (u, x). (4) The first order term S(u) describes spatial inhomogeneity and/or covariate effects. The higher order term V (u, x) describes interpoint interaction. 13/28

14 Fitting Gibbs models in spatstat The model with log-linear conditional intensity is fitted by calling ppm in the basic form ppm(x, ~ terms, V) The first argument X is the point pattern dataset and the second argument ~terms is a model formula, specifying the first order term S(u). The third argument V is an object of the special class "interact" which describes the interpoint interaction term V (u, x). Internally spatstat creates a quadrature scheme of class quad by: > Q <- quadscheme(x) Such an object simply contains two point patterns data and dummy and a vector of weights attached to every point. The quadrature scheme can be controlled by the user by making the desired data and dummy point patterns and then call: > Q <- quadscheme(data, dummy) 14/28

15 15/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Fitting Gibbs models in spatstat Examples of quadrature schemes: > U1 <- default.dummy(x, nd = 40) > Q1 <- quadscheme(x, U1) > U2 <- rpoint(1600, win = as.owin(x)) > Q2 <- quadscheme(x, U2, ntile = c(10, 10)) > plot(q1) > plot(q2) Q1 Q2

16 16/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Fitting Gibbs models in spatstat Plotting a fitted model generates image and contour plots of the fitted first order term exp(ˆη S(u)) the fitted conditional intensity λˆθ(u, x) > X <- swedishpines > fit <- ppm(x, ~polynom(x, y, 2), Strauss(r = 7)) > plot(fit, how = "image", ngrid = 256, pause = FALSE) Fitted trend Fitted cif

17 Fitting Gibbs models in spatstat For non-poisson models, it is also possible to extract and plot the interpoint interaction function, using fitin. > model <- ppm(x, ~1, PairPiece(seq(5, 25, by = 5))) > f <- fitin(model) > plot(f) Pairwise interaction h one Distance 17/28

18 18/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Simulation from fitted models A fitted Gibbs model can also be simulated automatically using rmh or simulate. > fit <- ppm(swedishpines, ~1, Strauss(r = 7)) > Xsim <- simulate(fit) > plot(xsim, main = "") Simulation 1

19 Simulation from fitted models The envelope command will also generate simulation envelopes for a fitted model. > plot(envelope(fit, nsim = 39), lwd = 2, main = "") K(r) obs mmean hi lo Spatial point processes: Theory and practice illustrated with r (one R unit = 0.1 metres) 19/28

20 Dealing with nuisance parameters Irregular parameters, such as the interaction radius r in the Strauss process, cannot be estimated directly using ppm. Indeed the statistical theory for estimating such parameters is unclear. For some special cases, a maximum likelihood estimator of the nuisance parameter is available. For example, for the hard core process with interaction radius r, the maximum likelihood estimator is the minimum nearest-neighbour distance. Thus the following is a reasonable approach to fit a hardcore process to a dataset X: > rhat <- min(nndist(x)) > rhat <- rhat * > ppm(x, ~1, Hardcore(hc = rhat)) 20/28

21 Dealing with nuisance parameters The analogue of profile likelihood, profile pseudolikelihood, provides a general solution which may or may not perform well. If θ = (φ, η) where φ denotes the nuisance parameters and η the regular parameters, define the profile log pseudolikelihood by PPL(φ, x) = max log PL ((φ, η); x). η The right hand side can be computed, for each fixed value of φ, by the algorithm ppm. Then we just have to maximise PPL(φ) over φ. This is done by the command profilepl: > df <- data.frame(r = seq(1, 15, by = 1)) > pfit <- profilepl(df, Strauss, swedishpines, ~1) 21/28

22 Dealing with nuisance parameters The profile pseudolikelihood can be plotted and the maximal value is indicated: > plot(pfit, main = "") /28

23 Improvements over maximum pseudolikelihood Currently the only pseudolikelihood alternative in spatstat is the Huang-Ogata one-step approximation to maximum likelihood. Starting from the maximum pseudolikelihood estimate ˆθ PL, we simulate M independent realisations of the model with parameters ˆθ PL, evaluate the canonical sufficient statistics, and use them to form estimates of the score and Fisher information at θ = ˆθ PL. Then we take one Newton-Raphson step, updating the value of θ. To use the Huang-Ogata method instead of maximum pseudolikelihood, add the argument method="ho". > fit <- ppm(swedishpines, ~1, Strauss(r = 10), + method = "ho") 23/28

24 Validation of fitted Gibbs models Suppose we want to validate a null model with density f θ (x). We can then consider the score test for goodness-of-fit againt an extended model: f θ,φ,r (x) = c θ,φ,r f θ (x) exp(φs(x, r)), where S(x, r) is an arbitrary summary statistic we wish to perturb the null model by. Notice that for fixed θ and r, f θ,φ,r (x) = c θ,φ,r f θ (x) exp(φs(x, r)), is a linear exponential family in φ. 24/28

25 Validation of fitted Gibbs models The score test of the null hypothesis H 0 : φ = 0 against the alternative hypothesis H 1 : φ 0, is based on the score test statistic T (θ, r) = U(0) = S(x, r) E (θ,0)[s(x, r)]. I (0) Var (θ,0) [S(X, r)] In practice θ is replaced by its estimate ˆθ under H 0 T = T (r) = S(x, r) E (ˆθ,0)[S(X, r)]. Var (ˆθ,0) [S(X, r)] So the score test for goodness-of-fit suggests that we should compare the functional summary statistic with its mean under the null model. 25/28

26 Residuals The Georgii-Nguyen-Zessin (GNZ) formula states that E [ ] s(x i, X \ {x i }) = E s(u, X)λ(u, X) du xi X R 2 for all measurable functions s such that the left or right hand side exists. Assume in the following that S is naturally expressible as a sum of local contributions S(x, r) = i s(x i, x i, r). (5) 26/28

27 Residuals Consider a null model with conditional intensity λ θ (u, x), and define the (s-weighted) innovation by I S(x, r) = S(x, r) s(u, x, r)λ θ (u, x) du (6) which by the GNZ formula has mean zero under the null model. In practice we replace θ by an estimate ˆθ (e.g. the MPLE) and consider the (s-weighted) residual R S(x, r) = S(x, r) s(u, x, r)λˆθ(u, x) du. (7) The residual shares many properties of the score function and can serve as a computationally efficient surrogate for the score. W W 27/28

28 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: Realization of an inhomogeneous Strauss process in the unit square showing strong inhibition and spatial trend with the number of points increasing from the left to the right hand side of the window

29 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: K^(r)

30 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: K^(r) (A): CK^(r)

31 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: K^(r) (A): CK^(r) (B): CK^(r)

32 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: K^(r) (A): CK^(r) (B): CK^(r) (C): CK^(r)

33 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: K^(r) (A): CK^(r) (B): CK^(r) (C): CK^(r) (D): CK^(r)

34 Residuals When S simply is the sum of points falling in B R 2 S(x, r) = i 1 {x i B} we obtain the raw residual R(B) = n(x B) B λ(u, x) du (9) where λ(u, x) is the conditional intensity of the fitted model, evaluated for the data point pattern x. If the fitted model is correct, the residuals have mean zero. If s(u, x) = 1/ λ(u, x) we obtain the Pearson residual R(B) = (x j x B) 1 λ(xj, x) B λ(u, x) du, which approximately has variance B. 29/28

35 Residuals When B(r) = {u W : Z(u) r} for a covariate function Z, we obtain a cummulative residual function. For example, for the raw residuals R(r) = R(B(r)) = n({x B(r)}) λ(u, x) du. B(r) This can be used to to check if we are missing a covariate in a fitted model. For example, consider a model for the rainforest data only containing the elevation as a covariate: > data(bei) > elev <- bei.extra$elev > slope <- bei.extra$grad > fit <- ppm(bei, ~elev, covariates = list(elev = elev)) 30/28

36 Residuals A plot of the Peason residuals as a function of r is called the lurking variable plot: > lurking(fit, slope, type = "pearson") /28

37 Thank you for your attention! 32/28

Advanced R course on spatial point patterns

Advanced R course on spatial point patterns Advanced R course on spatial point patterns Adrian Baddeley / Ege Rubak Curtin University / Aalborg University SSAI 2017 Spatial point patterns SSAI Course 2017 1 Plan of Workshop 1. Introduction 2. Inhomogeneous

More information

Normalising constants and maximum likelihood inference

Normalising constants and maximum likelihood inference Normalising constants and maximum likelihood inference Jakob G. Rasmussen Department of Mathematics Aalborg University Denmark March 9, 2011 1/14 Today Normalising constants Approximation of normalising

More information

Residual analysis for spatial point processes

Residual analysis for spatial point processes Residual analysis for spatial point processes A. addeley University of Western Australia, Perth, Australia R. Turner University of New runswick, Fredericton, Canada J. Møller University of Aalborg, Denmark

More information

Analysis of Gordon Square data for hybrids paper

Analysis of Gordon Square data for hybrids paper Analysis of Gordon Square data for hybrids paper Adrian Baddeley November 2, 2013 This Sweave script is part of the online supplementary material for the paper Hybrids of Gibbs point processes and their

More information

RESEARCH REPORT. A logistic regression estimating function for spatial Gibbs point processes.

RESEARCH REPORT. A logistic regression estimating function for spatial Gibbs point processes. CENTRE FOR STOCHASTIC GEOMETRY AND ADVANCED BIOIMAGING www.csgb.dk RESEARCH REPORT 2013 Adrian Baddeley, Jean-François Coeurjolly, Ege Rubak and Rasmus Waagepetersen A logistic regression estimating function

More information

Jesper Møller ) and Kateřina Helisová )

Jesper Møller ) and Kateřina Helisová ) Jesper Møller ) and ) ) Aalborg University (Denmark) ) Czech Technical University/Charles University in Prague 5 th May 2008 Outline 1. Describing model 2. Simulation 3. Power tessellation of a union of

More information

Estimating functions for inhomogeneous spatial point processes with incomplete covariate data

Estimating functions for inhomogeneous spatial point processes with incomplete covariate data Estimating functions for inhomogeneous spatial point processes with incomplete covariate data Rasmus aagepetersen Department of Mathematics Aalborg University Denmark August 15, 2007 1 / 23 Data (Barro

More information

Practical maximum pseudolikelihood. for spatial point patterns. Adrian Baddeley and Rolf Turner. Abstract

Practical maximum pseudolikelihood. for spatial point patterns. Adrian Baddeley and Rolf Turner. Abstract Practical maximum pseudolikelihood for spatial point patterns Adrian Baddeley and Rolf Turner Abstract We describe a technique for computing approximate maximum pseudolikelihood estimates of the parameters

More information

Modelling Spatial Point Patterns in R

Modelling Spatial Point Patterns in R Modelling Spatial Point Patterns in R Adrian Baddeley 1 and Rolf Turner 2 1 School of Mathematics and Statistics, University of Western Australia, Crawley 6009 WA, Australia, adrian@maths.uwa.edu.au 2

More information

Simulation of point process using MCMC

Simulation of point process using MCMC Simulation of point process using MCMC Jakob G. Rasmussen Department of Mathematics Aalborg University Denmark March 2, 2011 1/14 Today When do we need MCMC? Case 1: Fixed number of points Metropolis-Hastings,

More information

Gibbs point processes : modelling and inference

Gibbs point processes : modelling and inference Gibbs point processes : modelling and inference J.-F. Coeurjolly (Grenoble University) et J.-M Billiot, D. Dereudre, R. Drouilhet, F. Lavancier 03/09/2010 J.-F. Coeurjolly () Gibbs models 03/09/2010 1

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information

Lecture 2: Poisson point processes: properties and statistical inference

Lecture 2: Poisson point processes: properties and statistical inference Lecture 2: Poisson point processes: properties and statistical inference Jean-François Coeurjolly http://www-ljk.imag.fr/membres/jean-francois.coeurjolly/ 1 / 20 Definition, properties and simulation Statistical

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

Spatial point processes

Spatial point processes Mathematical sciences Chalmers University of Technology and University of Gothenburg Gothenburg, Sweden June 25, 2014 Definition A point process N is a stochastic mechanism or rule to produce point patterns

More information

Exponential families also behave nicely under conditioning. Specifically, suppose we write η = (η 1, η 2 ) R k R p k so that

Exponential families also behave nicely under conditioning. Specifically, suppose we write η = (η 1, η 2 ) R k R p k so that 1 More examples 1.1 Exponential families under conditioning Exponential families also behave nicely under conditioning. Specifically, suppose we write η = η 1, η 2 R k R p k so that dp η dm 0 = e ηt 1

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning

More information

Mathematical statistics

Mathematical statistics October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation

More information

Reminder of some Markov Chain properties:

Reminder of some Markov Chain properties: Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent

More information

ESTIMATING FUNCTIONS FOR INHOMOGENEOUS COX PROCESSES

ESTIMATING FUNCTIONS FOR INHOMOGENEOUS COX PROCESSES ESTIMATING FUNCTIONS FOR INHOMOGENEOUS COX PROCESSES Rasmus Waagepetersen Department of Mathematics, Aalborg University, Fredrik Bajersvej 7G, DK-9220 Aalborg, Denmark (rw@math.aau.dk) Abstract. Estimation

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation

More information

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Multilevel Statistical Models: 3 rd edition, 2003 Contents Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Stat 710: Mathematical Statistics Lecture 12

Stat 710: Mathematical Statistics Lecture 12 Stat 710: Mathematical Statistics Lecture 12 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 12 Feb 18, 2009 1 / 11 Lecture 12:

More information

Control Variates for Markov Chain Monte Carlo

Control Variates for Markov Chain Monte Carlo Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability

More information

Jean-Michel Billiot, Jean-François Coeurjolly and Rémy Drouilhet

Jean-Michel Billiot, Jean-François Coeurjolly and Rémy Drouilhet Colloque International de Statistique Appliquée pour le Développement en Afrique International Conference on Applied Statistics for Development in Africa Sada 07 nn, 1?? 007) MAXIMUM PSEUDO-LIKELIHOOD

More information

Aalborg Universitet. Properties of residuals for spatial point processes. Baddeley, A.; Møller, Jesper; Pakes, A.G. Publication date: 2006

Aalborg Universitet. Properties of residuals for spatial point processes. Baddeley, A.; Møller, Jesper; Pakes, A.G. Publication date: 2006 Downloaded from vbn.aau.dk on: pril 12, 219 alborg Universitet Properties of residuals for spatial point processes addeley,.; Møller, Jesper; Pakes,.G. Publication date: 26 Document Version Publisher's

More information

Connections between score matching, contrastive divergence, and pseudolikelihood for continuous-valued variables. Revised submission to IEEE TNN

Connections between score matching, contrastive divergence, and pseudolikelihood for continuous-valued variables. Revised submission to IEEE TNN Connections between score matching, contrastive divergence, and pseudolikelihood for continuous-valued variables Revised submission to IEEE TNN Aapo Hyvärinen Dept of Computer Science and HIIT University

More information

Bayesian Analysis of Spatial Point Patterns

Bayesian Analysis of Spatial Point Patterns Bayesian Analysis of Spatial Point Patterns by Thomas J. Leininger Department of Statistical Science Duke University Date: Approved: Alan E. Gelfand, Supervisor Robert L. Wolpert Merlise A. Clyde James

More information

Gibbs Voronoi tessellations: Modeling, Simulation, Estimation.

Gibbs Voronoi tessellations: Modeling, Simulation, Estimation. Gibbs Voronoi tessellations: Modeling, Simulation, Estimation. Frédéric Lavancier, Laboratoire Jean Leray, Nantes (France) Work with D. Dereudre (LAMAV, Valenciennes, France). SPA, Berlin, July 27-31,

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information

Modern statistics for spatial point processes

Modern statistics for spatial point processes Modern statistics for spatial point processes June 21, 2006 Jesper Møller and Rasmus P Waagepetersen Department of Mathematical Sciences, Aalborg University Abstract: We summarize and discuss the current

More information

Monte Carlo in Bayesian Statistics

Monte Carlo in Bayesian Statistics Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview

More information

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait

A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

MCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham

MCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham MCMC 2: Lecture 2 Coding and output Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham Contents 1. General (Markov) epidemic model 2. Non-Markov epidemic model 3. Debugging

More information

13 Notes on Markov Chain Monte Carlo

13 Notes on Markov Chain Monte Carlo 13 Notes on Markov Chain Monte Carlo Markov Chain Monte Carlo is a big, and currently very rapidly developing, subject in statistical computation. Many complex and multivariate types of random data, useful

More information

Markov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials)

Markov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials) Markov Networks l Like Bayes Nets l Graph model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov

More information

Fundamental Probability and Statistics

Fundamental Probability and Statistics Fundamental Probability and Statistics "There are known knowns. These are things we know that we know. There are known unknowns. That is to say, there are things that we know we don't know. But there are

More information

Aalborg Universitet. Modern statistics for spatial point processes Møller, Jesper; Waagepetersen, Rasmus Plenge. Publication date: 2006

Aalborg Universitet. Modern statistics for spatial point processes Møller, Jesper; Waagepetersen, Rasmus Plenge. Publication date: 2006 Aalborg Universitet Modern statistics for spatial point processes Møller, Jesper; Waagepetersen, Rasmus Plenge Publication date: 2006 Document Version Accepted author manuscript, peer reviewed version

More information

Answers and expectations

Answers and expectations Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E

More information

Statistical study of spatial dependences in long memory random fields on a lattice, point processes and random geometry.

Statistical study of spatial dependences in long memory random fields on a lattice, point processes and random geometry. Statistical study of spatial dependences in long memory random fields on a lattice, point processes and random geometry. Frédéric Lavancier, Laboratoire de Mathématiques Jean Leray, Nantes 9 décembre 2011.

More information

Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood

Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood PRE 906: Structural Equation Modeling Lecture #3 February 4, 2015 PRE 906, SEM: Estimation Today s Class An

More information

Markov Networks.

Markov Networks. Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Markov Chain Monte Carlo, Numerical Integration

Markov Chain Monte Carlo, Numerical Integration Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical

More information

Regression Models - Introduction

Regression Models - Introduction Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent

More information

Two step estimation for Neyman-Scott point process with inhomogeneous cluster centers. May 2012

Two step estimation for Neyman-Scott point process with inhomogeneous cluster centers. May 2012 Two step estimation for Neyman-Scott point process with inhomogeneous cluster centers Tomáš Mrkvička, Milan Muška, Jan Kubečka May 2012 Motivation Study of the influence of covariates on the occurrence

More information

A short introduction to INLA and R-INLA

A short introduction to INLA and R-INLA A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Theory and Methods of Statistical Inference

Theory and Methods of Statistical Inference PhD School in Statistics cycle XXIX, 2014 Theory and Methods of Statistical Inference Instructors: B. Liseo, L. Pace, A. Salvan (course coordinator), N. Sartori, A. Tancredi, L. Ventura Syllabus Some prerequisites:

More information

Markov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials)

Markov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials) Markov Networks l Like Bayes Nets l Graphical model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

Gibbs models estimation of galaxy point processes with the ABC Shadow algorithm

Gibbs models estimation of galaxy point processes with the ABC Shadow algorithm Gibbs models estimation of galaxy point processes with the ABC Shadow algorithm Lluı s Hurtado Gil Universidad CEU San Pablo (lluis.hurtadogil@ceu.es) Radu Stoica Universite de Lorraine, IECL (radu-stefan.stoica@univlorraine.fr)

More information

Analysing replicated point patterns in spatstat

Analysing replicated point patterns in spatstat Analysing replicated point patterns in spatstat Adrian Baddeley For spatstat version 1.57-1 Abstract This document describes spatstat s capabilities for fitting models to replicated point patterns. More

More information

Lecture 7 and 8: Markov Chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.

Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J. Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical

More information

Generalized, Linear, and Mixed Models

Generalized, Linear, and Mixed Models Generalized, Linear, and Mixed Models CHARLES E. McCULLOCH SHAYLER.SEARLE Departments of Statistical Science and Biometrics Cornell University A WILEY-INTERSCIENCE PUBLICATION JOHN WILEY & SONS, INC. New

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

Lecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1

Lecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1 Lecture 5 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,

More information

Petr Volf. Model for Difference of Two Series of Poisson-like Count Data

Petr Volf. Model for Difference of Two Series of Poisson-like Count Data Petr Volf Institute of Information Theory and Automation Academy of Sciences of the Czech Republic Pod vodárenskou věží 4, 182 8 Praha 8 e-mail: volf@utia.cas.cz Model for Difference of Two Series of Poisson-like

More information

The lmm Package. May 9, Description Some improved procedures for linear mixed models

The lmm Package. May 9, Description Some improved procedures for linear mixed models The lmm Package May 9, 2005 Version 0.3-4 Date 2005-5-9 Title Linear mixed models Author Original by Joseph L. Schafer . Maintainer Jing hua Zhao Description Some improved

More information

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown.

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown. Weighting We have seen that if E(Y) = Xβ and V (Y) = σ 2 G, where G is known, the model can be rewritten as a linear model. This is known as generalized least squares or, if G is diagonal, with trace(g)

More information

Package lmm. R topics documented: March 19, Version 0.4. Date Title Linear mixed models. Author Joseph L. Schafer

Package lmm. R topics documented: March 19, Version 0.4. Date Title Linear mixed models. Author Joseph L. Schafer Package lmm March 19, 2012 Version 0.4 Date 2012-3-19 Title Linear mixed models Author Joseph L. Schafer Maintainer Jing hua Zhao Depends R (>= 2.0.0) Description Some

More information

A Bayesian Approach to Phylogenetics

A Bayesian Approach to Phylogenetics A Bayesian Approach to Phylogenetics Niklas Wahlberg Based largely on slides by Paul Lewis (www.eeb.uconn.edu) An Introduction to Bayesian Phylogenetics Bayesian inference in general Markov chain Monte

More information

Alternative implementations of Monte Carlo EM algorithms for likelihood inferences

Alternative implementations of Monte Carlo EM algorithms for likelihood inferences Genet. Sel. Evol. 33 001) 443 45 443 INRA, EDP Sciences, 001 Alternative implementations of Monte Carlo EM algorithms for likelihood inferences Louis Alberto GARCÍA-CORTÉS a, Daniel SORENSEN b, Note a

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

ABC methods for phase-type distributions with applications in insurance risk problems

ABC methods for phase-type distributions with applications in insurance risk problems ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon

More information

Statistics & Data Sciences: First Year Prelim Exam May 2018

Statistics & Data Sciences: First Year Prelim Exam May 2018 Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book

More information

Approximate Bayesian Computation and Model Assessment for Repulsive Spatial Point Processes

Approximate Bayesian Computation and Model Assessment for Repulsive Spatial Point Processes Approximate Bayesian Computation and Model Assessment for Repulsive Spatial Point Processes arxiv:1604.07027v2 [stat.co] 26 Aug 2016 Shinichiro Shirota Department of Statistical Science, Duke University,

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 13-28 February 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Limitations of Gibbs sampling. Metropolis-Hastings algorithm. Proof

More information

Bayesian Inference in Astronomy & Astrophysics A Short Course

Bayesian Inference in Astronomy & Astrophysics A Short Course Bayesian Inference in Astronomy & Astrophysics A Short Course Tom Loredo Dept. of Astronomy, Cornell University p.1/37 Five Lectures Overview of Bayesian Inference From Gaussians to Periodograms Learning

More information

Hierarchical Linear Models

Hierarchical Linear Models Hierarchical Linear Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin The linear regression model Hierarchical Linear Models y N(Xβ, Σ y ) β σ 2 p(β σ 2 ) σ 2 p(σ 2 ) can be extended

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Nuisance parameters and their treatment

Nuisance parameters and their treatment BS2 Statistical Inference, Lecture 2, Hilary Term 2008 April 2, 2008 Ancillarity Inference principles Completeness A statistic A = a(x ) is said to be ancillary if (i) The distribution of A does not depend

More information

An introduction to spatial point processes

An introduction to spatial point processes An introduction to spatial point processes Jean-François Coeurjolly 1 Examples of spatial data 2 Intensities and Poisson p.p. 3 Summary statistics 4 Models for point processes A very very brief introduction...

More information

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do

More information

Maximum-Likelihood Estimation: Basic Ideas

Maximum-Likelihood Estimation: Basic Ideas Sociology 740 John Fox Lecture Notes Maximum-Likelihood Estimation: Basic Ideas Copyright 2014 by John Fox Maximum-Likelihood Estimation: Basic Ideas 1 I The method of maximum likelihood provides estimators

More information

Stochastic Proximal Gradient Algorithm

Stochastic Proximal Gradient Algorithm Stochastic Institut Mines-Télécom / Telecom ParisTech / Laboratoire Traitement et Communication de l Information Joint work with: Y. Atchade, Ann Arbor, USA, G. Fort LTCI/Télécom Paristech and the kind

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

STA216: Generalized Linear Models. Lecture 1. Review and Introduction

STA216: Generalized Linear Models. Lecture 1. Review and Introduction STA216: Generalized Linear Models Lecture 1. Review and Introduction Let y 1,..., y n denote n independent observations on a response Treat y i as a realization of a random variable Y i In the general

More information

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 71. Decide in each case whether the hypothesis is simple

More information

Chapter 3 : Likelihood function and inference

Chapter 3 : Likelihood function and inference Chapter 3 : Likelihood function and inference 4 Likelihood function and inference The likelihood Information and curvature Sufficiency and ancilarity Maximum likelihood estimation Non-regular models EM

More information

Contents. Part I: Fundamentals of Bayesian Inference 1

Contents. Part I: Fundamentals of Bayesian Inference 1 Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian

More information

SAMPLING ALGORITHMS. In general. Inference in Bayesian models

SAMPLING ALGORITHMS. In general. Inference in Bayesian models SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be

More information

Residuals and Goodness-of-fit tests for marked Gibbs point processes

Residuals and Goodness-of-fit tests for marked Gibbs point processes Residuals and Goodness-of-fit tests for marked Gibbs point processes Frédéric Lavancier (Laboratoire Jean Leray, Nantes, France) Joint work with J.-F. Coeurjolly (Grenoble, France) 09/06/2010 F. Lavancier

More information

ML estimation: Random-intercepts logistic model. and z

ML estimation: Random-intercepts logistic model. and z ML estimation: Random-intercepts logistic model log p ij 1 p = x ijβ + υ i with υ i N(0, συ) 2 ij Standardizing the random effect, θ i = υ i /σ υ, yields log p ij 1 p = x ij β + σ υθ i with θ i N(0, 1)

More information

Likelihood Inference for Lattice Spatial Processes

Likelihood Inference for Lattice Spatial Processes Likelihood Inference for Lattice Spatial Processes Donghoh Kim November 30, 2004 Donghoh Kim 1/24 Go to 1234567891011121314151617 FULL Lattice Processes Model : The Ising Model (1925), The Potts Model

More information

Data Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006

Data Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006 Astronomical p( y x, I) p( x, I) p ( x y, I) = p( y, I) Data Analysis I Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK 10 lectures, beginning October 2006 4. Monte Carlo Methods

More information

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017

Zig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017 Zig-Zag Monte Carlo Delft University of Technology Joris Bierkens February 7, 2017 Joris Bierkens (TU Delft) Zig-Zag Monte Carlo February 7, 2017 1 / 33 Acknowledgements Collaborators Andrew Duncan Paul

More information

Sampling Algorithms for Probabilistic Graphical models

Sampling Algorithms for Probabilistic Graphical models Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir

More information