Monte Carlo Solution of Integral Equations
|
|
- Robert Small
- 6 years ago
- Views:
Transcription
1 1 Monte Carlo Solution of Integral Equations of the second kind Adam M. Johansen www2.warwick.ac.uk/fac/sci/statistics/staff/academic/johansen/talks/ March 7th, 2011 MIRaW Monte Carlo Methods Day
2 2 Outline Background Importance Sampling IS Fredholm Equations of the 2nd Kind Sequential IS and 2nd Kind Fredholm Equations Methodology A Path-space Interpretation MCMC for 2nd Kind Fredholm Equations Results A Toy Example An Asset-Pricing Example
3 3 The Monte Carlo Method Consider estimating: I ϕ = E f [ϕx] = fxϕxdx Given X 1,..., X N iid f: lim N 1 N N ϕx i I ϕ i=1 Alteratively, approximate: and use fx = 1 N N δx x i 1=1 E f [ϕ X 1,..., X N ] E f [ϕ].
4 4 The Monte Carlo Method Consider estimating: I ϕ = E f [ϕx] = fxϕxdx Given X 1,..., X N iid f: lim N 1 N N ϕx i I ϕ i=1 Alteratively, approximate: and use fx = 1 N N δx x i 1=1 E f [ϕ X 1,..., X N ] E f [ϕ].
5 5 Importance Sampling If f g i.e. gx = 0 fx = 0: Iϕ = gx fx gx ϕxdx [ ] fx =E g ϕx gx }{{} wx Given X 1,..., X N iid g: lim N 1 N N wx i ϕx i I ϕ i=1 Alteratively, approximate: fx = 1 N N wx i δx X i 1=1 E f [ϕ X 1... X N ] E f [ϕ].
6 6 Importance Sampling If f g i.e. gx = 0 fx = 0: Iϕ = gx fx gx ϕxdx [ ] fx =E g ϕx gx }{{} wx Given X 1,..., X N iid g: lim N 1 N N wx i ϕx i I ϕ i=1 Alteratively, approximate: fx = 1 N N wx i δx X i 1=1 E f [ϕ X 1... X N ] E f [ϕ].
7 7 Fredholm Equations of the Second Kind Consider the equation: fx = E fykx, ydy + gx in which source gx and transition Kx, y are known, but fx is unknown. This is a Fredholm equation of the second kind. Finding fx is an inverse problem.
8 The Von Neumann Expansion We have the expansion: fx = fykx, ydy + gx E [ ] = fzky, zdx + gy Kx, ydy + gx E E. =gx + K n x, yfydy n=1 where: K 1 x, y =Kx, y K n x, y = E K n 1 x, zkz, ydz For convergence, it is sufficient that: gx + K n x, ygydy < n=1 8
9 9 A Sampling Approach Choose µ a pdf on E. Define E := E { } Let Mx, y be a Markov transition kernel on E such that: is absorbing Mx, = Pd and Mx, y > 0 whenever Kx, y > 0. The pair µ, M defines a Markov chain on E.
10 10 Sequential Importance Sampling SIS Algorithm Simulate N paths Calculate weights: W X i 0:k i = { } X i N until 0:k i Xi +1 i=1 k i +1 =. k i K X i 1 k 1,Xi k µ X i 0 k=1 M X i k 1,Xi k g X i 0 µ X i 0 g X i k i P d if k i 1, P d if k i = 0. Approximate fx 0 with: f x 0 = 1 N N W i=1 X i δ 0:k i x 0 X i 0
11 11 Von Neumann Representation Revisited Starting from fx 0 = gx 0 + K n x 0, x n fx n dx n n=1 E We have directly: n fx 0 =gx 0 + Kx p 1, x p gx n dx 1... dx n =f 0 x 0 + n=1 E n n=1 p=1 E n f n x 0:n dx 1... dx n with f n x 0:n := gx n n Kx p 1, x p. p=1
12 12 A Path-Space Interpretation of SIS Define: where πn, x 0:n =p n π n x 0 : n on F = k=0 {k} E k+1 p n =PX 0:n E n+1, X n+1 = = 1 P d n P d π n x 0:n = µx 0 n k=1 Mx k 1, x k 1 P d n Define also: Approximate with: fk, x 0:k =f 0 x 0 δ 0,k + f n x 0:n δ n,k n=1 ˆfx 0 = 1 N N i=1 fk i, X i 0:k i πk i, X i δx 0 X i 0 0:k i
13 13 Limitations of SIS Arbitary geometric distribution for k. Arbitary initial distribution. Importance weights 1 k i K µ M X i 0 k=1 X i k 1, Xi k X i k 1, Xi k g X i k i P d Resampling will not help.
14 14 Limitations of SIS Arbitary geometric distribution for k. Arbitary initial distribution. Importance weights 1 k i K µ M X i 0 k=1 X i k 1, Xi k X i k 1, Xi k g X i k i P d Resampling will not help.
15 15 Optimal Importance Distribution Minimize variance of absolute value of weights. Consider choosing πn, x 0:n = p k δ k,n π k x 0:k i=k π n x 0:n =c 1 n f n x 0:n c n = f n x 0:n dx 0:n E n+1 p n =c n /c c = c n. We have: n=0 f x 0 = c sgn f 0 x 0 π 0, x 0 + c sgn f n x 0:n π n, x 0:n dx 1:n. E n n=1
16 16 That s where the MCMC comes in Obtain a sample-approximation of π... using your favourite sampler, Reversible Jump MCMC 2, for instance Use the samples to approximate the optimal proposal. Combine with importance sampling. Estimate normalising constant, c.
17 17 A Simple RJMCMC Algorithm Initialization. Set k 1, X 1 randomly or deterministically. 0:k 1 Iteration i 2. Sample U U[0, 1] If U u k i 1 k i, x i Update Move. 0:k i Else If U u k i 1 + b k i 1: Else k i, x i 0:k i Birth Move. k i, x i 0:k i Death Move.
18 18 Very Simple Update Move Set k i = k i 1, sample J U { 0, 1,..., k i} and XJ q u X i 1 J,. With probability π k i, X i 1 min 1, 0:J 1, X J, Xi 1 π k i, X i 1 q 0:k i u set X i = X i 1 0:k i 0:J 1, X J, Xi 1, J+1:k i otherwise set X i 0:k i = X i 1 0:k i 1. Acceptance probabilities involve: π l, x 0:l π k, x 0:k = c lπ l x 0:l c k π k x 0:k = J+1:k i X i 1 J q u XJ, Xi 1 J, XJ f l x 0:l f k x 0:k.
19 19 Very Simple Update Move Set k i = k i 1, sample J U { 0, 1,..., k i} and XJ q u X i 1 J,. With probability π k i, X i 1 min 1, 0:J 1, X J, Xi 1 π k i, X i 1 q 0:k i u set X i = X i 1 0:k i 0:J 1, X J, Xi 1, J+1:k i otherwise set X i 0:k i = X i 1 0:k i 1. Acceptance probabilities involve: π l, x 0:l π k, x 0:k = c lπ l x 0:l c k π k x 0:k = J+1:k i X i 1 J q u XJ, Xi 1 J, XJ f l x 0:l f k x 0:k.
20 20 Very Simple Birth Move Sample J U { 0, 1,..., k i 1}, sample X J q b. With probability π k i 1 + 1, min 1, π X i 1 0:J 1, X J, Xi 1 d J:k i 1 k i 1 +1 k i 1, X i 1 0:k i 1 q b X J bk i 1 set k i = k i 1 + 1, X i 0:k = X i 1 0:J 1, X J, Xi 1, J:k i 1 otherwise set k i = k i 1, X i 0:k i = X i 1 0:k i 1.
21 21 Very Simple Death Move Sample J U { 0, 1,..., k i 1}. With probability π k i 1 1, min 1, π X i 1 0:J 1, Xi 1 J+1:k i 1 set k i = k i 1 1, X i 0:k i = q b k i 1, X i 1 d 0:k i 1 k i 1 X i 1 J X i 1 0:J 1, Xi 1, J+1:k i 1 otherwise set k i = k i 1, X i 0:k i = X i 1 0:k i 1. b k i 1 1
22 22 Estimating the Normalising Constant Estimate p n as: p n = 1 N Estimate directly, also, c 0 and hence n i=1 δ n,k i gxdx ĉ = c 0 / p 0 ĉ n =ĉ p n
23 23 Toy Example The simple algorithm was applied to fx = 1 The solution fx = expx. 0 1 expx y fydy + 2 } 3 {{} 3 expx. }{{} Kx,y gx Birth, death and update probabilities were set to 1/3. A uniform distribution over the unit interval was used for all proposals.
24 24 Point Estimation for the Toy Example 3.5 Point Estimates and their Standard Deviations Truth N=100 N=1,000 N=10,000 fx x 0 Fix x 0 and simulate everything else.
25 25 Estimating fx from Samples Given ˆfx = 1 N such that [ E N i=1 W i δx X i 0, ] ˆfxϕx = fxϕxdx Simple option: histogram ϕx = I A x. More subtle option: fx = Kx, yˆfydy + gx =gx + 1 N N i=1 W i Kx, X i 0.
26 26 Estimating fx from Samples Given ˆfx = 1 N such that [ E N i=1 W i δx X i 0, ] ˆfxϕx = fxϕxdx Simple option: histogram ϕx = I A x. More subtle option: fx = Kx, yˆfydy + gx =gx + 1 N N i=1 W i Kx, X i 0.
27 Estimated fx for the Toy Problem 3 Estimate 1 Estimate 2 Truth
28 28 An Asset-Pricing Problem The rational expectation pricing model cf. 4 requires: the price, V s of an asset in some state s E satisfies V s = πs + β V tpt sdt. E Where: πs denotes the return on investment, β is a discount factor and pt s is a Markov kernel which models the state evolution. Simple example: E =[0, 1] β =0.85 pt s = with a = 0.05, b = 0.85 and λ = 100. φ t as+b λ Φ 1 [as+b] λ Φ as+b λ
29 29 Estimated fx for asset pricing case
30 30 Simulated Chain Lengths for Different Starting Points 0.6 X00 X
31 31 Conclusions Many Monte Carlo methods can be viewed as importance sampling. Many problems can be recast as calculation of expectations. Approximate solution of Fredholm equations. Also applicable to Volterra Equations 3. Any Monte Carlo method could be used. Some room for further investigation of this method.
32 32 References [1] A. Doucet, A. M. Johansen, and V. B. Tadić. On solving integral equations using Markov Chain Monte Carlo. Applied Mathematics and Computation, 216: , [2] P. J. Green. Reversible jump Markov Chain Monte Carlo computation and Bayesian model determination. Biometrika, 82: , [3] G. W. Peters, A. M. Johansen, and A. Doucet. Simulation of the annual loss distribution in operational risk via Panjer recursions and Volterra integral equations for value at risk and expected shortfall estimation. Journal of Operational Risk, 23:29 58, Fall [4] J. Rust, J. F. Traub, and H. Wózniakowski. Is there a curse of dimensionality for contraction fixed points in the worst case? Econometrica, 701: , 2002.
On Solving Integral Equations using Markov Chain Monte Carlo Methods
On Solving Integral quations using Markov Chain Monte Carlo Methods Arnaud Doucet Departments of Statistics and Computer Science, University of British Columbia, Vancouver, BC, Canada Adam M. Johansen,
More informationAn Brief Overview of Particle Filtering
1 An Brief Overview of Particle Filtering Adam M. Johansen a.m.johansen@warwick.ac.uk www2.warwick.ac.uk/fac/sci/statistics/staff/academic/johansen/talks/ May 11th, 2010 Warwick University Centre for Systems
More informationExercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters
Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for
More informationMonte Carlo Approximation of Monte Carlo Filters
Monte Carlo Approximation of Monte Carlo Filters Adam M. Johansen et al. Collaborators Include: Arnaud Doucet, Axel Finke, Anthony Lee, Nick Whiteley 7th January 2014 Context & Outline Filtering in State-Space
More informationAuxiliary Particle Methods
Auxiliary Particle Methods Perspectives & Applications Adam M. Johansen 1 adam.johansen@bristol.ac.uk Oxford University Man Institute 29th May 2008 1 Collaborators include: Arnaud Doucet, Nick Whiteley
More informationAdvanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering
Advanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering Axel Gandy Department of Mathematics Imperial College London http://www2.imperial.ac.uk/~agandy London
More informationMarkov Chain Monte Carlo Methods for Stochastic Optimization
Markov Chain Monte Carlo Methods for Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U of Toronto, MIE,
More informationParticle Filtering Approaches for Dynamic Stochastic Optimization
Particle Filtering Approaches for Dynamic Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge I-Sim Workshop,
More informationParticle Filtering for Data-Driven Simulation and Optimization
Particle Filtering for Data-Driven Simulation and Optimization John R. Birge The University of Chicago Booth School of Business Includes joint work with Nicholas Polson. JRBirge INFORMS Phoenix, October
More informationLecture Particle Filters. Magnus Wiktorsson
Lecture Particle Filters Magnus Wiktorsson Monte Carlo filters The filter recursions could only be solved for HMMs and for linear, Gaussian models. Idea: Approximate any model with a HMM. Replace p(x)
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 18-16th March Arnaud Doucet
Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 18-16th March 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Trans-dimensional Markov chain Monte Carlo. Bayesian model for autoregressions.
More informationParticle Filters: Convergence Results and High Dimensions
Particle Filters: Convergence Results and High Dimensions Mark Coates mark.coates@mcgill.ca McGill University Department of Electrical and Computer Engineering Montreal, Quebec, Canada Bellairs 2012 Outline
More informationLecture 8: The Metropolis-Hastings Algorithm
30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:
More informationLecture Particle Filters
FMS161/MASM18 Financial Statistics November 29, 2010 Monte Carlo filters The filter recursions could only be solved for HMMs and for linear, Gaussian models. Idea: Approximate any model with a HMM. Replace
More informationApril 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning
for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions
More informationRobert Collins CSE586, PSU Intro to Sampling Methods
Intro to Sampling Methods CSE586 Computer Vision II Penn State Univ Topics to be Covered Monte Carlo Integration Sampling and Expected Values Inverse Transform Sampling (CDF) Ancestral Sampling Rejection
More informationMarkov Chain Monte Carlo Lecture 6
Sequential parallel tempering With the development of science and technology, we more and more need to deal with high dimensional systems. For example, we need to align a group of protein or DNA sequences
More informationThe Hierarchical Particle Filter
and Arnaud Doucet http://go.warwick.ac.uk/amjohansen/talks MCMSki V Lenzerheide 7th January 2016 Context & Outline Filtering in State-Space Models: SIR Particle Filters [GSS93] Block-Sampling Particle
More informationDivide-and-Conquer Sequential Monte Carlo
Divide-and-Conquer Joint work with: John Aston, Alexandre Bouchard-Côté, Brent Kirkpatrick, Fredrik Lindsten, Christian Næsseth, Thomas Schön University of Warwick a.m.johansen@warwick.ac.uk http://go.warwick.ac.uk/amjohansen/talks/
More informationMarkov Chain Monte Carlo Methods for Stochastic
Markov Chain Monte Carlo Methods for Stochastic Optimization i John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U Florida, Nov 2013
More informationMonte Carlo methods for sampling-based Stochastic Optimization
Monte Carlo methods for sampling-based Stochastic Optimization Gersende FORT LTCI CNRS & Telecom ParisTech Paris, France Joint works with B. Jourdain, T. Lelièvre, G. Stoltz from ENPC and E. Kuhn from
More informationSampling from complex probability distributions
Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.
Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Introduction to Markov chain Monte Carlo The Gibbs Sampler Examples Overview of the Lecture
More informationThree examples of a Practical Exact Markov Chain Sampling
Three examples of a Practical Exact Markov Chain Sampling Zdravko Botev November 2007 Abstract We present three examples of exact sampling from complex multidimensional densities using Markov Chain theory
More informationFunctional Analysis HW 2
Brandon Behring Functional Analysis HW 2 Exercise 2.6 The space C[a, b] equipped with the L norm defined by f = b a f(x) dx is incomplete. If f n f with respect to the sup-norm then f n f with respect
More informationControlled sequential Monte Carlo
Controlled sequential Monte Carlo Jeremy Heng, Department of Statistics, Harvard University Joint work with Adrian Bishop (UTS, CSIRO), George Deligiannidis & Arnaud Doucet (Oxford) Bayesian Computation
More informationStat 516, Homework 1
Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball
More informationA Note on Auxiliary Particle Filters
A Note on Auxiliary Particle Filters Adam M. Johansen a,, Arnaud Doucet b a Department of Mathematics, University of Bristol, UK b Departments of Statistics & Computer Science, University of British Columbia,
More informationON A CHARACTERIZATION OF THE KERNEL OF THE DIRICHLET-TO-NEUMANN MAP FOR A PLANAR REGION
ON A CHARACTERIZATION OF THE KERNEL OF THE DIRICHLET-TO-NEUMANN MAP FOR A PLANAR REGION DAVID INGERMAN AND JAMES A. MORROW Abstract. We will show the Dirichlet-to-Neumann map Λ for the electrical conductivity
More informationLikelihood-free MCMC
Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte
More informationSequential Monte Carlo and Particle Filtering. Frank Wood Gatsby, November 2007
Sequential Monte Carlo and Particle Filtering Frank Wood Gatsby, November 2007 Importance Sampling Recall: Let s say that we want to compute some expectation (integral) E p [f] = p(x)f(x)dx and we remember
More information( ) ( ) Monte Carlo Methods Interested in. E f X = f x d x. Examples:
Monte Carlo Methods Interested in Examples: µ E f X = f x d x Type I error rate of a hypothesis test Mean width of a confidence interval procedure Evaluating a likelihood Finding posterior mean and variance
More information2.2 Separable Equations
2.2 Separable Equations Definition A first-order differential equation that can be written in the form Is said to be separable. Note: the variables of a separable equation can be written as Examples Solve
More informationRare Event Simulation using Importance Sampling and Cross-Entropy p. 1
Rare Event Simulation using Importance Sampling and Cross-Entropy Caitlin James Supervisor: Phil Pollett Rare Event Simulation using Importance Sampling and Cross-Entropy p. 1 500 Simulation of a population
More informationRobert Collins CSE586, PSU Intro to Sampling Methods
Intro to Sampling Methods CSE586 Computer Vision II Penn State Univ Topics to be Covered Monte Carlo Integration Sampling and Expected Values Inverse Transform Sampling (CDF) Ancestral Sampling Rejection
More informationBoundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta
Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta November 29, 2007 Outline Overview of Kernel Density
More informationSequential Monte Carlo Methods for Bayesian Model Selection in Positron Emission Tomography
Methods for Bayesian Model Selection in Positron Emission Tomography Yan Zhou John A.D. Aston and Adam M. Johansen 6th January 2014 Y. Zhou J. A. D. Aston and A. M. Johansen Outline Positron emission tomography
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationScientific Computing: Monte Carlo
Scientific Computing: Monte Carlo Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 April 5th and 12th, 2012 A. Donev (Courant Institute)
More informationIntegral approximation by kernel smoothing
Integral approximation by kernel smoothing François Portier Université catholique de Louvain - ISBA August, 29 2014 In collaboration with Bernard Delyon Topic of the talk: Given ϕ : R d R, estimation of
More informationOn Markov chain Monte Carlo methods for tall data
On Markov chain Monte Carlo methods for tall data Remi Bardenet, Arnaud Doucet, Chris Holmes Paper review by: David Carlson October 29, 2016 Introduction Many data sets in machine learning and computational
More informationHmms with variable dimension structures and extensions
Hmm days/enst/january 21, 2002 1 Hmms with variable dimension structures and extensions Christian P. Robert Université Paris Dauphine www.ceremade.dauphine.fr/ xian Hmm days/enst/january 21, 2002 2 1 Estimating
More informationF denotes cumulative density. denotes probability density function; (.)
BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models
More informationParticle Filtering a brief introductory tutorial. Frank Wood Gatsby, August 2007
Particle Filtering a brief introductory tutorial Frank Wood Gatsby, August 2007 Problem: Target Tracking A ballistic projectile has been launched in our direction and may or may not land near enough to
More informationA Review of Pseudo-Marginal Markov Chain Monte Carlo
A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the
More informationAn introduction to particle filters
An introduction to particle filters Andreas Svensson Department of Information Technology Uppsala University June 10, 2014 June 10, 2014, 1 / 16 Andreas Svensson - An introduction to particle filters Outline
More informationLearning Static Parameters in Stochastic Processes
Learning Static Parameters in Stochastic Processes Bharath Ramsundar December 14, 2012 1 Introduction Consider a Markovian stochastic process X T evolving (perhaps nonlinearly) over time variable T. We
More informationSmoothing Algorithms for the Probability Hypothesis Density Filter
Smoothing Algorithms for the Probability Hypothesis Density Filter Sergio Hernández Laboratorio de Procesamiento de Información GeoEspacial. Universidad Católica del Maule. Talca, Chile. shernandez@ucm.cl.
More informationKernel Sequential Monte Carlo
Kernel Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) * equal contribution April 25, 2016 1 / 37 Section
More informationL09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms
L09. PARTICLE FILTERING NA568 Mobile Robotics: Methods & Algorithms Particle Filters Different approach to state estimation Instead of parametric description of state (and uncertainty), use a set of state
More informationNegative Association, Ordering and Convergence of Resampling Methods
Negative Association, Ordering and Convergence of Resampling Methods Nicolas Chopin ENSAE, Paristech (Joint work with Mathieu Gerber and Nick Whiteley, University of Bristol) Resampling schemes: Informal
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods
Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationSupplementary Notes: Segment Parameter Labelling in MCMC Change Detection
Supplementary Notes: Segment Parameter Labelling in MCMC Change Detection Alireza Ahrabian 1 arxiv:1901.0545v1 [eess.sp] 16 Jan 019 Abstract This work addresses the problem of segmentation in time series
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo
Group Prof. Daniel Cremers 11. Sampling Methods: Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative
More informationParticle Filters. Outline
Particle Filters M. Sami Fadali Professor of EE University of Nevada Outline Monte Carlo integration. Particle filter. Importance sampling. Degeneracy Resampling Example. 1 2 Monte Carlo Integration Numerical
More informationBayesian Inference for Dirichlet-Multinomials
Bayesian Inference for Dirichlet-Multinomials Mark Johnson Macquarie University Sydney, Australia MLSS Summer School 1 / 50 Random variables and distributed according to notation A probability distribution
More informationThe Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision
The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that
More informationDynamic Risk Measures and Nonlinear Expectations with Markov Chain noise
Dynamic Risk Measures and Nonlinear Expectations with Markov Chain noise Robert J. Elliott 1 Samuel N. Cohen 2 1 Department of Commerce, University of South Australia 2 Mathematical Insitute, University
More informationRobert Collins CSE586, PSU Intro to Sampling Methods
Robert Collins Intro to Sampling Methods CSE586 Computer Vision II Penn State Univ Robert Collins A Brief Overview of Sampling Monte Carlo Integration Sampling and Expected Values Inverse Transform Sampling
More informationIntroduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak
Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,
More informationTowards a Bayesian model for Cyber Security
Towards a Bayesian model for Cyber Security Mark Briers (mbriers@turing.ac.uk) Joint work with Henry Clausen and Prof. Niall Adams (Imperial College London) 27 September 2017 The Alan Turing Institute
More informationAn introduction to Sequential Monte Carlo
An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods
More informationErgodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.
Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions
More informationCOMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017
COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University FEATURE EXPANSIONS FEATURE EXPANSIONS
More informationKernel adaptive Sequential Monte Carlo
Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline
More informationSequential Monte Carlo Samplers for Applications in High Dimensions
Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex
More informationSequential Monte Carlo samplers
J. R. Statist. Soc. B (2006) 68, Part 3, pp. 411 436 Sequential Monte Carlo samplers Pierre Del Moral, Université Nice Sophia Antipolis, France Arnaud Doucet University of British Columbia, Vancouver,
More informationData Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006
Astronomical p( y x, I) p( x, I) p ( x y, I) = p( y, I) Data Analysis I Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK 10 lectures, beginning October 2006 4. Monte Carlo Methods
More informationS6880 #13. Variance Reduction Methods
S6880 #13 Variance Reduction Methods 1 Variance Reduction Methods Variance Reduction Methods 2 Importance Sampling Importance Sampling 3 Control Variates Control Variates Cauchy Example Revisited 4 Antithetic
More informationAsymptotics and Simulation of Heavy-Tailed Processes
Asymptotics and Simulation of Heavy-Tailed Processes Department of Mathematics Stockholm, Sweden Workshop on Heavy-tailed Distributions and Extreme Value Theory ISI Kolkata January 14-17, 2013 Outline
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend
More informationBlind Equalization via Particle Filtering
Blind Equalization via Particle Filtering Yuki Yoshida, Kazunori Hayashi, Hideaki Sakai Department of System Science, Graduate School of Informatics, Kyoto University Historical Remarks A sequential Monte
More informationIntroduction to Markov Chain Monte Carlo & Gibbs Sampling
Introduction to Markov Chain Monte Carlo & Gibbs Sampling Prof. Nicholas Zabaras Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Ithaca, NY 14853-3801 Email: zabaras@cornell.edu
More informationComputer Vision Group Prof. Daniel Cremers. 14. Sampling Methods
Prof. Daniel Cremers 14. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationBayesian Methods with Monte Carlo Markov Chains II
Bayesian Methods with Monte Carlo Markov Chains II Henry Horng-Shing Lu Institute of Statistics National Chiao Tung University hslu@stat.nctu.edu.tw http://tigpbp.iis.sinica.edu.tw/courses.htm 1 Part 3
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationPartially Collapsed Gibbs Samplers: Theory and Methods. Ever increasing computational power along with ever more sophisticated statistical computing
Partially Collapsed Gibbs Samplers: Theory and Methods David A. van Dyk 1 and Taeyoung Park Ever increasing computational power along with ever more sophisticated statistical computing techniques is making
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet
Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 13-28 February 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Limitations of Gibbs sampling. Metropolis-Hastings algorithm. Proof
More informationWinter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo
Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 9: Markov Chain Monte Carlo 9.1 Markov Chain A Markov Chain Monte
More informationRecent Developments in Auxiliary Particle Filtering
Chapter 3 Recent Developments in Auxiliary Particle Filtering Nick Whiteley 1 and Adam M. Johansen 2 3.1 Background 3.1.1 State Space Models State space models (SSMs; sometimes termed hidden Markov models,
More informationA Comparison of the EKF, SPKF, and the Bayes Filter for Landmark-Based Localization
A Comparison of the EKF, SPKF, and the Bayes Filter for Landmark-Based Localization and Timothy D. Barfoot CRV 2 Outline Background Objective Experimental Setup Results Discussion Conclusion 2 Outline
More informationChapter 7. Markov chain background. 7.1 Finite state space
Chapter 7 Markov chain background A stochastic process is a family of random variables {X t } indexed by a varaible t which we will think of as time. Time can be discrete or continuous. We will only consider
More informationAssessing Regime Uncertainty Through Reversible Jump McMC
Assessing Regime Uncertainty Through Reversible Jump McMC August 14, 2008 1 Introduction Background Research Question 2 The RJMcMC Method McMC RJMcMC Algorithm Dependent Proposals Independent Proposals
More informationOn the flexibility of Metropolis-Hastings acceptance probabilities in auxiliary variable proposal generation
On the flexibility of Metropolis-Hastings acceptance probabilities in auxiliary variable proposal generation Geir Storvik Department of Mathematics and (sfi) 2 Statistics for Innovation, University of
More informationIntroduction to Bayesian methods in inverse problems
Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction
More informationLangevin and Hamiltonian based Sequential MCMC for Efficient Bayesian Filtering in High-dimensional Spaces
1 Langevin and Hamiltonian based Sequential MCMC for Efficient Bayesian Filtering in High-dimensional Spaces arxiv:154.5715v [stat.co] 9 Oct 15 François Septier, Member, IEEE, and Gareth W. Peters Accepted
More information6 Markov Chain Monte Carlo (MCMC)
6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution
More information1: PROBABILITY REVIEW
1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following
More informationBits of Machine Learning Part 1: Supervised Learning
Bits of Machine Learning Part 1: Supervised Learning Alexandre Proutiere and Vahan Petrosyan KTH (The Royal Institute of Technology) Outline of the Course 1. Supervised Learning Regression and Classification
More informationAn ABC interpretation of the multiple auxiliary variable method
School of Mathematical and Physical Sciences Department of Mathematics and Statistics Preprint MPS-2016-07 27 April 2016 An ABC interpretation of the multiple auxiliary variable method by Dennis Prangle
More informationMarkov chain Monte Carlo
1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop
More informationIntroduction to Curve Estimation
Introduction to Curve Estimation Density 0.000 0.002 0.004 0.006 700 800 900 1000 1100 1200 1300 Wilcoxon score Michael E. Tarter & Micheal D. Lock Model-Free Curve Estimation Monographs on Statistics
More informationThree hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.
Three hours To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER EXTREME VALUES AND FINANCIAL RISK Examiner: Answer QUESTION 1, QUESTION
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of
More informationBayesian Graphical Models
Graphical Models and Inference, Lecture 16, Michaelmas Term 2009 December 4, 2009 Parameter θ, data X = x, likelihood L(θ x) p(x θ). Express knowledge about θ through prior distribution π on θ. Inference
More informationCalibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods
Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods Jonas Hallgren 1 1 Department of Mathematics KTH Royal Institute of Technology Stockholm, Sweden BFS 2012 June
More informationTSRT14: Sensor Fusion Lecture 8
TSRT14: Sensor Fusion Lecture 8 Particle filter theory Marginalized particle filter Gustaf Hendeby gustaf.hendeby@liu.se TSRT14 Lecture 8 Gustaf Hendeby Spring 2018 1 / 25 Le 8: particle filter theory,
More informationComputer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo
Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain
More informationBrief introduction to Markov Chain Monte Carlo
Brief introduction to Department of Probability and Mathematical Statistics seminar Stochastic modeling in economics and finance November 7, 2011 Brief introduction to Content 1 and motivation Classical
More information