Spatial point processes: Theory and practice illustrated with R
|
|
- Alisha Morgan
- 5 years ago
- Views:
Transcription
1 Spatial point processes: Theory and practice illustrated with R Department of Mathematical Sciences Aalborg University Lecture IV, February 24, /28
2 Contents of lecture IV Papangelou conditional intensity. Simulating Gibbs models. Maximum pseudolikelihood. Fitting Gibbs models. Residuals and model validation. 2/28
3 Papangelou conditional intensity The main tool for analysing a Gibbs point process is the Papangelou conditional intensity λ(u, X). Informally, the conditional probability of finding a point of the process inside an infinitesimal neighbourhood of the location u, given the complete point pattern at all other locations, is λ(u, X) du. For point processes in a bounded window, the conditional intensity at a location u given the configuration x is related to the probability density f by λ(u, x) = f (x {u}) f (x) (for u x), the ratio of the probability densities for the configuration x with and without the point u added. 3/28
4 Papangelou conditional intensity The Poisson process with intensity function λ(u) has conditional intensity λ(u, x) = λ(u). The conditional intensity for a Poisson process does not depend on the configuration x, because the points of a Poisson process are independent. For the general pairwise interaction process the conditional intensity is n(x) λ(u, x) = b(u) c(u, x i ). i=1 4/28
5 Papangelou conditional intensity For the hard core process, { β if u xi > r for all i λ(u, x) = 0 otherwise which has the nice interpretation that a point u is either permitted or not permitted depending on whether it satisfies the hard core requirement. For the Strauss process λ(u, x) = βγ t(u,x) where t(u, x) = s(x {u}) s(x) is the number of points of x that lie within a distance r of the location u. For γ < 1, this has the interpretation that a random point is less likely to occur at the location u if there are many points in the neighbourhood. (1) 5/28
6 Papangelou conditional intensity For the area-interaction process, λ(u, x) = βγ B(u,x) where B(u, x) = A(x {u}) A(x) is the area of that part of the disc of radius r centred on u that is not covered by discs of radius r centred at the other points x i x. Strauss area interaction + + 6/28
7 Simulating Gibbs models Gibbs models can be simulated by Markov chain Monte Carlo algorithms. (Actually, MCMC algorithms were invented to simulate Gibbs processes.) Currently spatstat offers the function rmh which simulates Gibbs processes using the Metropolis-Hastings algorithm. > rmh(model, start, control) model determines the point process model to be simulated (see help(rmhmodel)). start determines the initial state of the Markov chain (see help(rmhstart)). control specifies control parameters for running the Markov chain, such as the number of iteration steps (see help(rmhcontrol)). 7/28
8 Simulating Gibbs models In the simplest uses of rmh, the three arguments are lists of parameter values. To generate a simulated realisation of the Strauss process with parameters β = 2, γ = 0.7, r = 0.7 in a square of side 10, > mo <- list(cif = "strauss", par = list(beta = 2, + gamma = 0.2, r = 0.7), w = square(10)) > X <- rmh(model = mo, start = list(n.start = 42), + control = list(nrep = 1e+06)) The other arguments specify a random initial state of 42 points, and that the algorithm shall be run for a million iterations. 8/28
9 9/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Simulating Gibbs models Specifically for the Strauss model spatstat offers the possibility of using a perfect simulation algorithm to generate a realisation from the model: > Y <- rstrauss(beta = 2, gamma = 0.2, R = 0.7, + W = square(10)) > plot(x, main = "") > plot(y, main = "")
10 Maximum pseudolikelihood Maximum likelihood estimation is in general very difficult for point process models, and an alternative is to maximise the log pseudolikelihood log PL (θ; x) = log λ(x i ; x) λ(u, x) du. (2) i W In general it is not a likelihood, but the analogue of the score equation log PL (θ) = 0 θ is an unbiased estimating equation. Thus the maximum pseudolikelihood estimator is asymptotically unbiased, consistent and asymptotically normal under appropriate conditions. 10/28
11 Maximum pseudolikelihood Consider any quadrature scheme to approximate the integral m λ(u, x) du λ θ (u j, x)w j W where u j, j = 1,..., m, are points in W and w j > 0 are quadrature weights summing to W. If the quadrature points are chosen such that {x 1,..., x n } {u 1,..., u m } we can write m log PL (θ; x) = (y j log λ j λ j )w j, j=1 j=1 where λ j = λ θ (u j, x) and y j = z j /w j, and { 1 if uj is a data point, u z j = j x 0 if u j is a dummy point, u j / x. This is the log-likelihood of independent Poisson variables Y j with means λ j taken with weights w j. 11/28
12 Fitting Gibbs models in spatstat The function ppm implements maximum pseudolikelihood inference for general Gibbs processes based on the conditional intensity of the model, λ θ (u, x). The model must be loglinear in the parameters θ: log λ θ (u, x) = θ S(u, x), (3) where S(u, x) is a real-valued or vector-valued function of location u and configuration x. Parameters θ appearing in the loglinear form (2) are called regular parameters, and all other parameters are irregular parameters. 12/28
13 Fitting Gibbs models in spatstat For example, the Strauss process conditional intensity can be recast as log λ(u, x) = log β + (log γ)t(u, x) so that θ = (log β, log γ) are regular parameters, but the interaction distance r is an irregular parameter. In spatstat the conditional intensity is split into first-order and higher-order terms: log λ θ (u, x) = η S(u) + ϕ V (u, x). (4) The first order term S(u) describes spatial inhomogeneity and/or covariate effects. The higher order term V (u, x) describes interpoint interaction. 13/28
14 Fitting Gibbs models in spatstat The model with log-linear conditional intensity is fitted by calling ppm in the basic form ppm(x, ~ terms, V) The first argument X is the point pattern dataset and the second argument ~terms is a model formula, specifying the first order term S(u). The third argument V is an object of the special class "interact" which describes the interpoint interaction term V (u, x). Internally spatstat creates a quadrature scheme of class quad by: > Q <- quadscheme(x) Such an object simply contains two point patterns data and dummy and a vector of weights attached to every point. The quadrature scheme can be controlled by the user by making the desired data and dummy point patterns and then call: > Q <- quadscheme(data, dummy) 14/28
15 15/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Fitting Gibbs models in spatstat Examples of quadrature schemes: > U1 <- default.dummy(x, nd = 40) > Q1 <- quadscheme(x, U1) > U2 <- rpoint(1600, win = as.owin(x)) > Q2 <- quadscheme(x, U2, ntile = c(10, 10)) > plot(q1) > plot(q2) Q1 Q2
16 16/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Fitting Gibbs models in spatstat Plotting a fitted model generates image and contour plots of the fitted first order term exp(ˆη S(u)) the fitted conditional intensity λˆθ(u, x) > X <- swedishpines > fit <- ppm(x, ~polynom(x, y, 2), Strauss(r = 7)) > plot(fit, how = "image", ngrid = 256, pause = FALSE) Fitted trend Fitted cif
17 Fitting Gibbs models in spatstat For non-poisson models, it is also possible to extract and plot the interpoint interaction function, using fitin. > model <- ppm(x, ~1, PairPiece(seq(5, 25, by = 5))) > f <- fitin(model) > plot(f) Pairwise interaction h one Distance 17/28
18 18/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Simulation from fitted models A fitted Gibbs model can also be simulated automatically using rmh or simulate. > fit <- ppm(swedishpines, ~1, Strauss(r = 7)) > Xsim <- simulate(fit) > plot(xsim, main = "") Simulation 1
19 Simulation from fitted models The envelope command will also generate simulation envelopes for a fitted model. > plot(envelope(fit, nsim = 39), lwd = 2, main = "") K(r) obs mmean hi lo Spatial point processes: Theory and practice illustrated with r (one R unit = 0.1 metres) 19/28
20 Dealing with nuisance parameters Irregular parameters, such as the interaction radius r in the Strauss process, cannot be estimated directly using ppm. Indeed the statistical theory for estimating such parameters is unclear. For some special cases, a maximum likelihood estimator of the nuisance parameter is available. For example, for the hard core process with interaction radius r, the maximum likelihood estimator is the minimum nearest-neighbour distance. Thus the following is a reasonable approach to fit a hardcore process to a dataset X: > rhat <- min(nndist(x)) > rhat <- rhat * > ppm(x, ~1, Hardcore(hc = rhat)) 20/28
21 Dealing with nuisance parameters The analogue of profile likelihood, profile pseudolikelihood, provides a general solution which may or may not perform well. If θ = (φ, η) where φ denotes the nuisance parameters and η the regular parameters, define the profile log pseudolikelihood by PPL(φ, x) = max log PL ((φ, η); x). η The right hand side can be computed, for each fixed value of φ, by the algorithm ppm. Then we just have to maximise PPL(φ) over φ. This is done by the command profilepl: > df <- data.frame(r = seq(1, 15, by = 1)) > pfit <- profilepl(df, Strauss, swedishpines, ~1) 21/28
22 Dealing with nuisance parameters The profile pseudolikelihood can be plotted and the maximal value is indicated: > plot(pfit, main = "") /28
23 Improvements over maximum pseudolikelihood Currently the only pseudolikelihood alternative in spatstat is the Huang-Ogata one-step approximation to maximum likelihood. Starting from the maximum pseudolikelihood estimate ˆθ PL, we simulate M independent realisations of the model with parameters ˆθ PL, evaluate the canonical sufficient statistics, and use them to form estimates of the score and Fisher information at θ = ˆθ PL. Then we take one Newton-Raphson step, updating the value of θ. To use the Huang-Ogata method instead of maximum pseudolikelihood, add the argument method="ho". > fit <- ppm(swedishpines, ~1, Strauss(r = 10), + method = "ho") 23/28
24 Validation of fitted Gibbs models Suppose we want to validate a null model with density f θ (x). We can then consider the score test for goodness-of-fit againt an extended model: f θ,φ,r (x) = c θ,φ,r f θ (x) exp(φs(x, r)), where S(x, r) is an arbitrary summary statistic we wish to perturb the null model by. Notice that for fixed θ and r, f θ,φ,r (x) = c θ,φ,r f θ (x) exp(φs(x, r)), is a linear exponential family in φ. 24/28
25 Validation of fitted Gibbs models The score test of the null hypothesis H 0 : φ = 0 against the alternative hypothesis H 1 : φ 0, is based on the score test statistic T (θ, r) = U(0) = S(x, r) E (θ,0)[s(x, r)]. I (0) Var (θ,0) [S(X, r)] In practice θ is replaced by its estimate ˆθ under H 0 T = T (r) = S(x, r) E (ˆθ,0)[S(X, r)]. Var (ˆθ,0) [S(X, r)] So the score test for goodness-of-fit suggests that we should compare the functional summary statistic with its mean under the null model. 25/28
26 Residuals The Georgii-Nguyen-Zessin (GNZ) formula states that E [ ] s(x i, X \ {x i }) = E s(u, X)λ(u, X) du xi X R 2 for all measurable functions s such that the left or right hand side exists. Assume in the following that S is naturally expressible as a sum of local contributions S(x, r) = i s(x i, x i, r). (5) 26/28
27 Residuals Consider a null model with conditional intensity λ θ (u, x), and define the (s-weighted) innovation by I S(x, r) = S(x, r) s(u, x, r)λ θ (u, x) du (6) which by the GNZ formula has mean zero under the null model. In practice we replace θ by an estimate ˆθ (e.g. the MPLE) and consider the (s-weighted) residual R S(x, r) = S(x, r) s(u, x, r)λˆθ(u, x) du. (7) The residual shares many properties of the score function and can serve as a computationally efficient surrogate for the score. W W 27/28
28 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: Realization of an inhomogeneous Strauss process in the unit square showing strong inhibition and spatial trend with the number of points increasing from the left to the right hand side of the window
29 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: K^(r)
30 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: K^(r) (A): CK^(r)
31 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: K^(r) (A): CK^(r) (B): CK^(r)
32 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: K^(r) (A): CK^(r) (B): CK^(r) (C): CK^(r)
33 28/28 Simulation Maximum pseudolikelihood Fitting Gibbs models Model validation Residuals The data-dependent integral C S(x, r) = W s(u, x, r)λˆθ (u, x) du (8) is the compensator of S. Example for different null models: K^(r) (A): CK^(r) (B): CK^(r) (C): CK^(r) (D): CK^(r)
34 Residuals When S simply is the sum of points falling in B R 2 S(x, r) = i 1 {x i B} we obtain the raw residual R(B) = n(x B) B λ(u, x) du (9) where λ(u, x) is the conditional intensity of the fitted model, evaluated for the data point pattern x. If the fitted model is correct, the residuals have mean zero. If s(u, x) = 1/ λ(u, x) we obtain the Pearson residual R(B) = (x j x B) 1 λ(xj, x) B λ(u, x) du, which approximately has variance B. 29/28
35 Residuals When B(r) = {u W : Z(u) r} for a covariate function Z, we obtain a cummulative residual function. For example, for the raw residuals R(r) = R(B(r)) = n({x B(r)}) λ(u, x) du. B(r) This can be used to to check if we are missing a covariate in a fitted model. For example, consider a model for the rainforest data only containing the elevation as a covariate: > data(bei) > elev <- bei.extra$elev > slope <- bei.extra$grad > fit <- ppm(bei, ~elev, covariates = list(elev = elev)) 30/28
36 Residuals A plot of the Peason residuals as a function of r is called the lurking variable plot: > lurking(fit, slope, type = "pearson") /28
37 Thank you for your attention! 32/28
Advanced R course on spatial point patterns
Advanced R course on spatial point patterns Adrian Baddeley / Ege Rubak Curtin University / Aalborg University SSAI 2017 Spatial point patterns SSAI Course 2017 1 Plan of Workshop 1. Introduction 2. Inhomogeneous
More informationNormalising constants and maximum likelihood inference
Normalising constants and maximum likelihood inference Jakob G. Rasmussen Department of Mathematics Aalborg University Denmark March 9, 2011 1/14 Today Normalising constants Approximation of normalising
More informationResidual analysis for spatial point processes
Residual analysis for spatial point processes A. addeley University of Western Australia, Perth, Australia R. Turner University of New runswick, Fredericton, Canada J. Møller University of Aalborg, Denmark
More informationAnalysis of Gordon Square data for hybrids paper
Analysis of Gordon Square data for hybrids paper Adrian Baddeley November 2, 2013 This Sweave script is part of the online supplementary material for the paper Hybrids of Gibbs point processes and their
More informationRESEARCH REPORT. A logistic regression estimating function for spatial Gibbs point processes.
CENTRE FOR STOCHASTIC GEOMETRY AND ADVANCED BIOIMAGING www.csgb.dk RESEARCH REPORT 2013 Adrian Baddeley, Jean-François Coeurjolly, Ege Rubak and Rasmus Waagepetersen A logistic regression estimating function
More informationJesper Møller ) and Kateřina Helisová )
Jesper Møller ) and ) ) Aalborg University (Denmark) ) Czech Technical University/Charles University in Prague 5 th May 2008 Outline 1. Describing model 2. Simulation 3. Power tessellation of a union of
More informationEstimating functions for inhomogeneous spatial point processes with incomplete covariate data
Estimating functions for inhomogeneous spatial point processes with incomplete covariate data Rasmus aagepetersen Department of Mathematics Aalborg University Denmark August 15, 2007 1 / 23 Data (Barro
More informationPractical maximum pseudolikelihood. for spatial point patterns. Adrian Baddeley and Rolf Turner. Abstract
Practical maximum pseudolikelihood for spatial point patterns Adrian Baddeley and Rolf Turner Abstract We describe a technique for computing approximate maximum pseudolikelihood estimates of the parameters
More informationModelling Spatial Point Patterns in R
Modelling Spatial Point Patterns in R Adrian Baddeley 1 and Rolf Turner 2 1 School of Mathematics and Statistics, University of Western Australia, Crawley 6009 WA, Australia, adrian@maths.uwa.edu.au 2
More informationSimulation of point process using MCMC
Simulation of point process using MCMC Jakob G. Rasmussen Department of Mathematics Aalborg University Denmark March 2, 2011 1/14 Today When do we need MCMC? Case 1: Fixed number of points Metropolis-Hastings,
More informationGibbs point processes : modelling and inference
Gibbs point processes : modelling and inference J.-F. Coeurjolly (Grenoble University) et J.-M Billiot, D. Dereudre, R. Drouilhet, F. Lavancier 03/09/2010 J.-F. Coeurjolly () Gibbs models 03/09/2010 1
More informationLecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011
Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector
More informationLecture 2: Poisson point processes: properties and statistical inference
Lecture 2: Poisson point processes: properties and statistical inference Jean-François Coeurjolly http://www-ljk.imag.fr/membres/jean-francois.coeurjolly/ 1 / 20 Definition, properties and simulation Statistical
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationSpatial point processes
Mathematical sciences Chalmers University of Technology and University of Gothenburg Gothenburg, Sweden June 25, 2014 Definition A point process N is a stochastic mechanism or rule to produce point patterns
More informationExponential families also behave nicely under conditioning. Specifically, suppose we write η = (η 1, η 2 ) R k R p k so that
1 More examples 1.1 Exponential families under conditioning Exponential families also behave nicely under conditioning. Specifically, suppose we write η = η 1, η 2 R k R p k so that dp η dm 0 = e ηt 1
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More informationReminder of some Markov Chain properties:
Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent
More informationESTIMATING FUNCTIONS FOR INHOMOGENEOUS COX PROCESSES
ESTIMATING FUNCTIONS FOR INHOMOGENEOUS COX PROCESSES Rasmus Waagepetersen Department of Mathematics, Aalborg University, Fredrik Bajersvej 7G, DK-9220 Aalborg, Denmark (rw@math.aau.dk) Abstract. Estimation
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationParameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn
Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation
More informationMultilevel Statistical Models: 3 rd edition, 2003 Contents
Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationStat 710: Mathematical Statistics Lecture 12
Stat 710: Mathematical Statistics Lecture 12 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 12 Feb 18, 2009 1 / 11 Lecture 12:
More informationControl Variates for Markov Chain Monte Carlo
Control Variates for Markov Chain Monte Carlo Dellaportas, P., Kontoyiannis, I., and Tsourti, Z. Dept of Statistics, AUEB Dept of Informatics, AUEB 1st Greek Stochastics Meeting Monte Carlo: Probability
More informationJean-Michel Billiot, Jean-François Coeurjolly and Rémy Drouilhet
Colloque International de Statistique Appliquée pour le Développement en Afrique International Conference on Applied Statistics for Development in Africa Sada 07 nn, 1?? 007) MAXIMUM PSEUDO-LIKELIHOOD
More informationAalborg Universitet. Properties of residuals for spatial point processes. Baddeley, A.; Møller, Jesper; Pakes, A.G. Publication date: 2006
Downloaded from vbn.aau.dk on: pril 12, 219 alborg Universitet Properties of residuals for spatial point processes addeley,.; Møller, Jesper; Pakes,.G. Publication date: 26 Document Version Publisher's
More informationConnections between score matching, contrastive divergence, and pseudolikelihood for continuous-valued variables. Revised submission to IEEE TNN
Connections between score matching, contrastive divergence, and pseudolikelihood for continuous-valued variables Revised submission to IEEE TNN Aapo Hyvärinen Dept of Computer Science and HIIT University
More informationBayesian Analysis of Spatial Point Patterns
Bayesian Analysis of Spatial Point Patterns by Thomas J. Leininger Department of Statistical Science Duke University Date: Approved: Alan E. Gelfand, Supervisor Robert L. Wolpert Merlise A. Clyde James
More informationGibbs Voronoi tessellations: Modeling, Simulation, Estimation.
Gibbs Voronoi tessellations: Modeling, Simulation, Estimation. Frédéric Lavancier, Laboratoire Jean Leray, Nantes (France) Work with D. Dereudre (LAMAV, Valenciennes, France). SPA, Berlin, July 27-31,
More informationEM Algorithm II. September 11, 2018
EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data
More informationModern statistics for spatial point processes
Modern statistics for spatial point processes June 21, 2006 Jesper Møller and Rasmus P Waagepetersen Department of Mathematical Sciences, Aalborg University Abstract: We summarize and discuss the current
More informationMonte Carlo in Bayesian Statistics
Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview
More informationA Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait
A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute
More informationComputer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo
Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationMCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham
MCMC 2: Lecture 2 Coding and output Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham Contents 1. General (Markov) epidemic model 2. Non-Markov epidemic model 3. Debugging
More information13 Notes on Markov Chain Monte Carlo
13 Notes on Markov Chain Monte Carlo Markov Chain Monte Carlo is a big, and currently very rapidly developing, subject in statistical computation. Many complex and multivariate types of random data, useful
More informationMarkov Networks. l Like Bayes Nets. l Graph model that describes joint probability distribution using tables (AKA potentials)
Markov Networks l Like Bayes Nets l Graph model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov
More informationFundamental Probability and Statistics
Fundamental Probability and Statistics "There are known knowns. These are things we know that we know. There are known unknowns. That is to say, there are things that we know we don't know. But there are
More informationAalborg Universitet. Modern statistics for spatial point processes Møller, Jesper; Waagepetersen, Rasmus Plenge. Publication date: 2006
Aalborg Universitet Modern statistics for spatial point processes Møller, Jesper; Waagepetersen, Rasmus Plenge Publication date: 2006 Document Version Accepted author manuscript, peer reviewed version
More informationAnswers and expectations
Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E
More informationStatistical study of spatial dependences in long memory random fields on a lattice, point processes and random geometry.
Statistical study of spatial dependences in long memory random fields on a lattice, point processes and random geometry. Frédéric Lavancier, Laboratoire de Mathématiques Jean Leray, Nantes 9 décembre 2011.
More informationMaximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood
Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood PRE 906: Structural Equation Modeling Lecture #3 February 4, 2015 PRE 906, SEM: Estimation Today s Class An
More informationMarkov Networks.
Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationMarkov Chain Monte Carlo, Numerical Integration
Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical
More informationRegression Models - Introduction
Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent
More informationTwo step estimation for Neyman-Scott point process with inhomogeneous cluster centers. May 2012
Two step estimation for Neyman-Scott point process with inhomogeneous cluster centers Tomáš Mrkvička, Milan Muška, Jan Kubečka May 2012 Motivation Study of the influence of covariates on the occurrence
More informationA short introduction to INLA and R-INLA
A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationTheory and Methods of Statistical Inference
PhD School in Statistics cycle XXIX, 2014 Theory and Methods of Statistical Inference Instructors: B. Liseo, L. Pace, A. Salvan (course coordinator), N. Sartori, A. Tancredi, L. Ventura Syllabus Some prerequisites:
More informationMarkov Networks. l Like Bayes Nets. l Graphical model that describes joint probability distribution using tables (AKA potentials)
Markov Networks l Like Bayes Nets l Graphical model that describes joint probability distribution using tables (AKA potentials) l Nodes are random variables l Labels are outcomes over the variables Markov
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationGibbs models estimation of galaxy point processes with the ABC Shadow algorithm
Gibbs models estimation of galaxy point processes with the ABC Shadow algorithm Lluı s Hurtado Gil Universidad CEU San Pablo (lluis.hurtadogil@ceu.es) Radu Stoica Universite de Lorraine, IECL (radu-stefan.stoica@univlorraine.fr)
More informationAnalysing replicated point patterns in spatstat
Analysing replicated point patterns in spatstat Adrian Baddeley For spatstat version 1.57-1 Abstract This document describes spatstat s capabilities for fitting models to replicated point patterns. More
More informationLecture 7 and 8: Markov Chain Monte Carlo
Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationMS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari
MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth
More informationLecture 8: Information Theory and Statistics
Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang
More informationDeblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.
Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical
More informationGeneralized, Linear, and Mixed Models
Generalized, Linear, and Mixed Models CHARLES E. McCULLOCH SHAYLER.SEARLE Departments of Statistical Science and Biometrics Cornell University A WILEY-INTERSCIENCE PUBLICATION JOHN WILEY & SONS, INC. New
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationLecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1
Lecture 5 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,
More informationPetr Volf. Model for Difference of Two Series of Poisson-like Count Data
Petr Volf Institute of Information Theory and Automation Academy of Sciences of the Czech Republic Pod vodárenskou věží 4, 182 8 Praha 8 e-mail: volf@utia.cas.cz Model for Difference of Two Series of Poisson-like
More informationThe lmm Package. May 9, Description Some improved procedures for linear mixed models
The lmm Package May 9, 2005 Version 0.3-4 Date 2005-5-9 Title Linear mixed models Author Original by Joseph L. Schafer . Maintainer Jing hua Zhao Description Some improved
More informationNow consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown.
Weighting We have seen that if E(Y) = Xβ and V (Y) = σ 2 G, where G is known, the model can be rewritten as a linear model. This is known as generalized least squares or, if G is diagonal, with trace(g)
More informationPackage lmm. R topics documented: March 19, Version 0.4. Date Title Linear mixed models. Author Joseph L. Schafer
Package lmm March 19, 2012 Version 0.4 Date 2012-3-19 Title Linear mixed models Author Joseph L. Schafer Maintainer Jing hua Zhao Depends R (>= 2.0.0) Description Some
More informationA Bayesian Approach to Phylogenetics
A Bayesian Approach to Phylogenetics Niklas Wahlberg Based largely on slides by Paul Lewis (www.eeb.uconn.edu) An Introduction to Bayesian Phylogenetics Bayesian inference in general Markov chain Monte
More informationAlternative implementations of Monte Carlo EM algorithms for likelihood inferences
Genet. Sel. Evol. 33 001) 443 45 443 INRA, EDP Sciences, 001 Alternative implementations of Monte Carlo EM algorithms for likelihood inferences Louis Alberto GARCÍA-CORTÉS a, Daniel SORENSEN b, Note a
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationABC methods for phase-type distributions with applications in insurance risk problems
ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon
More informationStatistics & Data Sciences: First Year Prelim Exam May 2018
Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book
More informationApproximate Bayesian Computation and Model Assessment for Repulsive Spatial Point Processes
Approximate Bayesian Computation and Model Assessment for Repulsive Spatial Point Processes arxiv:1604.07027v2 [stat.co] 26 Aug 2016 Shinichiro Shirota Department of Statistical Science, Duke University,
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet
Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 13-28 February 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Limitations of Gibbs sampling. Metropolis-Hastings algorithm. Proof
More informationBayesian Inference in Astronomy & Astrophysics A Short Course
Bayesian Inference in Astronomy & Astrophysics A Short Course Tom Loredo Dept. of Astronomy, Cornell University p.1/37 Five Lectures Overview of Bayesian Inference From Gaussians to Periodograms Learning
More informationHierarchical Linear Models
Hierarchical Linear Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin The linear regression model Hierarchical Linear Models y N(Xβ, Σ y ) β σ 2 p(β σ 2 ) σ 2 p(σ 2 ) can be extended
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationNuisance parameters and their treatment
BS2 Statistical Inference, Lecture 2, Hilary Term 2008 April 2, 2008 Ancillarity Inference principles Completeness A statistic A = a(x ) is said to be ancillary if (i) The distribution of A does not depend
More informationAn introduction to spatial point processes
An introduction to spatial point processes Jean-François Coeurjolly 1 Examples of spatial data 2 Intensities and Poisson p.p. 3 Summary statistics 4 Models for point processes A very very brief introduction...
More informationMultilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2
Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do
More informationMaximum-Likelihood Estimation: Basic Ideas
Sociology 740 John Fox Lecture Notes Maximum-Likelihood Estimation: Basic Ideas Copyright 2014 by John Fox Maximum-Likelihood Estimation: Basic Ideas 1 I The method of maximum likelihood provides estimators
More informationStochastic Proximal Gradient Algorithm
Stochastic Institut Mines-Télécom / Telecom ParisTech / Laboratoire Traitement et Communication de l Information Joint work with: Y. Atchade, Ann Arbor, USA, G. Fort LTCI/Télécom Paristech and the kind
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationSTA216: Generalized Linear Models. Lecture 1. Review and Introduction
STA216: Generalized Linear Models Lecture 1. Review and Introduction Let y 1,..., y n denote n independent observations on a response Treat y i as a realization of a random variable Y i In the general
More informationThe University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80
The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 71. Decide in each case whether the hypothesis is simple
More informationChapter 3 : Likelihood function and inference
Chapter 3 : Likelihood function and inference 4 Likelihood function and inference The likelihood Information and curvature Sufficiency and ancilarity Maximum likelihood estimation Non-regular models EM
More informationContents. Part I: Fundamentals of Bayesian Inference 1
Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian
More informationSAMPLING ALGORITHMS. In general. Inference in Bayesian models
SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be
More informationResiduals and Goodness-of-fit tests for marked Gibbs point processes
Residuals and Goodness-of-fit tests for marked Gibbs point processes Frédéric Lavancier (Laboratoire Jean Leray, Nantes, France) Joint work with J.-F. Coeurjolly (Grenoble, France) 09/06/2010 F. Lavancier
More informationML estimation: Random-intercepts logistic model. and z
ML estimation: Random-intercepts logistic model log p ij 1 p = x ijβ + υ i with υ i N(0, συ) 2 ij Standardizing the random effect, θ i = υ i /σ υ, yields log p ij 1 p = x ij β + σ υθ i with θ i N(0, 1)
More informationLikelihood Inference for Lattice Spatial Processes
Likelihood Inference for Lattice Spatial Processes Donghoh Kim November 30, 2004 Donghoh Kim 1/24 Go to 1234567891011121314151617 FULL Lattice Processes Model : The Ising Model (1925), The Potts Model
More informationData Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006
Astronomical p( y x, I) p( x, I) p ( x y, I) = p( y, I) Data Analysis I Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK 10 lectures, beginning October 2006 4. Monte Carlo Methods
More informationZig-Zag Monte Carlo. Delft University of Technology. Joris Bierkens February 7, 2017
Zig-Zag Monte Carlo Delft University of Technology Joris Bierkens February 7, 2017 Joris Bierkens (TU Delft) Zig-Zag Monte Carlo February 7, 2017 1 / 33 Acknowledgements Collaborators Andrew Duncan Paul
More informationSampling Algorithms for Probabilistic Graphical models
Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir
More information