Particle Filtering a brief introductory tutorial. Frank Wood Gatsby, August 2007
|
|
- Ralf Price
- 5 years ago
- Views:
Transcription
1 Particle Filtering a brief introductory tutorial Frank Wood Gatsby, August 2007
2 Problem: Target Tracking A ballistic projectile has been launched in our direction and may or may not land near enough to kill us We have orders to launch an interceptor only if there is a >5% chance it will land near enough to kill us (the projectile s kill radius is 100 meters) the interceptor requires 5 seconds to destroy incoming projectiles We live under an evil dictator who will kill us himself if we unnecessarily launch an interceptor (they re expensive after all)
3 Problem Schematic (x t, y t ) r t θ t (0,0)
4 Problem problems Only noisy observations are available True trajectory is unknown and must be inferred from the noisy observations Full details will be given at the end of the lecture
5 Probabilistic approach Treat true trajectory as a sequence of latent variables Specify a model and do inference to recover the distribution over the latent variables (trajectories)
6 Linear State Space Model (LSSM) Discrete time First-order Markov chain x t+1 = ax t + ², ² N(µ ², σ² 2 ) y t = bx t + η, η N(µ η, ση) 2 x t 1 x t x t+1 y t 1 y t y t+1
7 LSSM Joint distribution p(x 1:i, y 1:i )= Q N i=1 p(y i x i )p(x i x i 1 ) For prediction we want the posterior predictive p(x i y 1:i 1 ) and posterior (filtering) distributions p(x i 1 y 1:i 1 )
8 Inferring the distributions of interest Many methods exist to infer these distributions Markov Chain Monte Carlo (MCMC) Variational inference Belief propagation etc. In this setting sequential inference is possible because of characteristics of the model structure and preferable due to the problem requirements
9 Exploiting LSSM model structure p(x i y 1:i ) = p(x i, x 1:i 1 y 1:i )dx 1:i 1 p(y i x i, x 1:i 1, y 1:i 1 )p(x i, x 1:i 1, y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x 1:i 1, y 1:i )p(x 1:i 1 y 1:i )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x 1:i 1 y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 p(y i x i )p(x i y 1:i 1 )
10 Exploiting LSSM model structure p(x i y 1:i ) = Use p(xbayes i, x 1:i 1 rule yand 1:i the )dxconditional 1:i 1 independence structure dictated by the LSSM graphical model p(y i x i, x 1:i 1, y 1:i 1 )p(x i, x 1:i 1, y 1:i 1 )dx 1:i 1 x t 1 x t x t+1 p(y i x i ) p(x i x 1:i 1, y 1:i )p(x 1:i 1 y 1:i )dx 1:i 1 p(y i x y t 1 i ) p(x i y x t i 1 )p(x y t+1 1:i 1 y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 p(y i x i )p(x i y 1:i 1 )
11 for sequential inference p(x i y 1:i ) = p(x i, x 1:i 1 y 1:i )dx 1:i 1 p(y i x i, x 1:i 1, y 1:i 1 )p(x i, x 1:i 1, y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x 1:i 1, y 1:i )p(x 1:i 1 y 1:i )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x 1:i 1 y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 p(y i x i )p(x i y 1:i 1 )
12 Particle filtering steps Start with a discrete representation of the posterior up to observation i-1 Use Monte Carlo integration to represent the posterior predictive distribution as a finite mixture model Use importance sampling with the posterior predictive distribution as the proposal distribution to sample the posterior distribution up to observation i
13 Posterior predictive distribution What? State model Likelihood p(x i y 1:i ) p(y i x i ) p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 p(y i x i )p(x i y 1:i 1 ) Start with a discrete representation of this distribution
14 Monte Carlo integration General setup p(x) LX `=1 w`δ x` lim L LX `=1 w`f(x`) = f(x)p(x)dx As applied in this stage of the particle filter p(x i y 1:i 1 ) = p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 LX `=1 w i 1 ` p(x i x i 1 ` ) Samples from Finite mixture model
15 Intuitions Particle filter name comes from physical interpretation of samples and the fact that the filtering distribution is often the distribution of interest
16 Start with samples representing the hidden state {w i 1 `,x i 1 ` } L`=1 p(x i 1 y 1:i 1 ) (0,0)
17 Evolve them according to the state model ˆx i m p( x i 1 ` ) ŵ i m w i 1 m (0,0)
18 Re-weight them by the likelihood w i m ŵ i mp(y i ˆx i m) (0,0)
19 Down-sample / resample {w ì,x ì } L`=1 {wi m,x i m} M m=1 (0,0)
20 Particle filtering Consists of two basic elements: Monte Carlo integration LX p(x) w`δ x` lim L LX `=1 w`f(x`) = Importance sampling `=1 f(x)p(x)dx
21 Importance sampling E x [f(x)] = Original distribution: hard to sample from, easy to evaluate = Importance weights p(x)f(x)dx p(x) q(x) f(x)q(x)dx 1 LX p(x`) L q(x`) f(x`) `=1 r` = p(x`) q(x`) Proposal distribution: easy to sample from x` q( )
22 Importance sampling un-normalized distributions p(x) = p(x) p q(x) = q(x) q Un-normalized distribution to sample from, still hard to sample from and easy to evaluate x` q( ) Un-normalized proposal distribution: still easy to sample from New term: ratio of normalizing constants E x [f(x)] 1 L LX `=1 q p 1 L p(x`) q(x`) f(x`) LX `=1 p(x`) q(x`) f(x`)
23 Normalizing the importance weights Un-normalized importance weights Takes a little algebra r` = p(x`) q(x`) q L p P L `=1 r` w` = r` P L `=1 r` Normalized importance weights E x [f(x)] q p 1 L LX `=1 LX `=1 w`f(x`) p(x`) q(x`) f(x`)
24 PF ing: Forming the posterior predictive {w i 1 `,x i 1 ` } L`=1 p(x i 1 y 1:i 1 ) p(x i y 1:i 1 ) = Posterior up to observation i 1 p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 LX `=1 w i 1 ` p(x i x i 1 ` ) The proposal distribution for importance sampling of the posterior up to observation i is this approximate posterior predictive distribution
25 Sampling the posterior predictive Generating samples from the posterior predictive distribution is the first place where we can introduce variance reduction techniques {ŵ i m, ˆxi m }M m=1 p(x i y 1:i 1 ), p(x i y 1:i 1 ) For instance sample from each mixture component several twice such that M, the number of samples drawn, is two times L, the number of densities in the mixture model, and assign weights ŵ i m = wi 1 ` 2 LX `=1 w i 1 ` p(x i x i 1 ` )
26 Not the best Most efficient Monte Carlo estimator of a function Γ(x) From survey sampling theory: Neyman allocation Number drawn from each mixture density is proportional to the weight of the mixture density times the std. dev. of the function Γ over the mixture density Take home: smarter sampling possible [Cochran 1963]
27 Over-sampling from the posterior predictive distribution {ŵ i m, ˆx i m} M m=1 p(x i y 1:i 1 ) (0,0)
28 Importance sampling the posterior Recall that we want samples from p(x i y 1:i ) p(y i x i ) p(y i x i )p(x i y 1:i 1 ) p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 and make the following importance sampling identifications p(x i )=p(y i x i )p(x i y 1:i 1 ) q(x i ) = p(x i y 1:i 1 ) LX Distribution from which we want to sample Proposal distribution `=1 w i 1 ` p(x i x i 1 ` )
29 Sequential importance sampling Weighted posterior samples arise as {ŵ i m, ˆx i m} q( ) Normalization of the weights takes place as before wm i ˆr = P ì L `=1 ˆrì ˆr i m = p(ˆxi m ) q(ˆx i m ) = p(y i ˆx i m)ŵ i m We are left with M weighted samples from the posterior up to observation i {w i m,x i m} M m=1 p(x i y 1:i )
30 An alternative view {ŵ i m, ˆxi m } L X `=1 w i 1 ` p(x i x i 1 ` ) p(x i y 1:i 1 ) P M m=1 ŵi mδˆx i m p(x i y 1:i ) P M m=1 p(y i ˆx i m)ŵ i mδˆx i m
31 Importance sampling from the posterior distribution {w i m,x i m} M m=1 p(x i y 1:i ) (0,0)
32 Sequential importance re-sampling Down-sample L particles and weights from the collection of M particles and weights {w ì,x ì } L`=1 {wi m,x i m} M m=1 this can be done via multinomial sampling or in a way that provably minimizes estimator variance [Fearnhead 04]
33 Down-sampling the particle set {w ì,x ì } L`=1 p(x i y 1:i ) (0,0)
34 Recap Starting with (weighted) samples from the posterior up to observation i-1 Monte Carlo integration was used to form a mixture model representation of the posterior predictive distribution The posterior predictive distribution was used as a proposal distribution for importance sampling of the posterior up to observation i M>Lsamples were drawn and re-weighted according to the likelihood (the importance weight), then the collection of particles was down-sampled to Lweighted samples
35 SIR Particle Filter The Sequential Importance Sampler (SIS) particle filter multiplicatively updates weights at every iteration and never resamples particles (skipped) Concentrated weight / particle degeneracy The Sequential Importance Resampler (SIR) particle filter avoids many of the problems associated with SIS particle filtering but always uses representations with Lparticles and does not maintain a weighted particle set
36 Particle evolution {w i 1 `,x i 1 ` } L`=1 p(x i 1 y 1:i 1 ) (0,0)
37 Particle evolution {ŵ i m, ˆx i m} M m=1 p(x i y 1:i 1 ) (0,0)
38 Particle evolution {w i m,x i m} M m=1 p(x i y 1:i ) (0,0)
39 Particle evolution {w ì,x ì } L`=1 p(x i y 1:i ) (0,0)
40 LSSM Not alone Various other models are amenable to sequential inference, Dirichlet process mixture modelling is another example, dynamic Bayes nets are another
41 Tricks and Variants Reduce the dimensionality of the integrand through analytic integration Rao-Blackwellization Reduce the variance of the Monte Carlo estimator through Maintaining a weighted particle set Stratified sampling Over-sampling Optimal re-sampling
42 Rao-Blackwellization In models where parameters can be analytically marginalized out, or the particle state space can otherwise be collapsed, the efficiency of the particle filter can be improved by doing so
43 Stratified Sampling Sampling from a mixture density using the algorithm on the right produces a more efficient Monte Carlo estimator {w n,x n } K n=1 X k π k f k ( ) for n=1:k choose kaccording to π k sample x n from f k set w n equal to 1/N for n=1:k choose f n sample x n from f n set w n equal to π n
44 Intuition: weighted particle set What is the difference between these two discrete distributions over the set {a,b,c}? (a), (a), (b), (b), (c) (.4, a), (.4, b), (.2, c) Weighted particle representations are equally or more efficient for the same number of particles
45 Optimal particle re-weighting Next step: when down-sampling, pass all particles above a threshold c through without modifying their weights where c is the unique solution to P M m=1 min{cwi m, 1} = L Resample all particles with weights below c using stratified sampling and give the selected particles weight 1/ c Fearnhead 2004
46 Result is provably optimal In the down-sampling step {w ì,x ì } L`=1 {wi m,x i m} M m=1 Imagine instead a sparse set of weights of which some are zero { w ì,x ì } M`=1 {wi m,x i m} M m=1 Then this down-sampling algorithm is optimal w.r.t. P M m=1 E w[( w i m w i m) 2 ] Fearnhead 2004
47 Problem Details Time and position are given in seconds and meters respectively Initial launch velocity and position are both unknown The maximum muzzle velocity of the projectile is 1000m/s The measurement error in the Cartesian coordinate system is N(0,10000) and N(0,500) for x and y position respectively The measurement error in the polar coordinate system is N(0,.001) for θand Gamma(1,100) for r The kill radius of the projectile is 100m
48 Data and Support Code
49 Laws of Motion In case you ve forgotten: r =(v 0 cos(α))ti +((v 0 sin(α)t 1 2 gt2 )j where v is the initial speed and α is the initial angle
50 Good luck!
51 Monte Carlo Integration Compute integrals for which analytical solutions are unknown f(x)p(x)dx p(x) LX w`δ x` `=1
52 Monte Carlo Integration Integral approximated as the weighted sum of function evaluations at L points f(x)p(x)dx p(x) LX w`δ x` `=1 = LX `=1 f(x) LX `=1 w`f(x`) w`δ x`dx
53 Sampling To use MC integration one must be able to sample from p(x) lim L {w`,x`} L`=1 p( ) LX w`δ x` p( ) `=1
54 Theory (Convergence) Quality of the approximation independent of the dimensionality of the integrand Convergence of integral result to the truth is O(1/n 1/2 ) from L.L.N. s. Error is independent of dimensionality of x
55 Bayesian Modelling Formal framework for the expression of modelling assumptions Two common definitions: using Bayes rule marginalizing over models Posterior Prior p(θ x) = p(x θ)p(θ) p(x) = p(x θ)p(θ) p(x θ)p(θ)dθ Likelihood Evidence
56 Posterior Estimation Often the distribution over latent random variables (parameters) is of interest Sometimes this is easy (conjugacy) Usually it is hard because computing the evidence is intractable
57 Conjugacy Example θ Beta(α, β) x θ Binomial(N,θ) p(θ) = p(x θ) = Γ(α + β) Γ(α)Γ(β) θα 1 (1 θ) β 1 µ N θ x (1 θ) N x x x successes in N trials, θ probability of success
58 Conjugacy Continued p(θ x) = = = p(x θ)p(θ) R p(x θ)p(θ)dθ 1 (x) p(x θ)p(θ) 1 (x) θα 1+x (1 θ) β 1+N x θ x Beta(α + x, β + N x) µ Γ(α + β + N) (x) = Γ(α + x)γ(β + N x) 1
59 Non-Conjugate Models Easy to come up with examples σ 2 N(0, α) x σ 2 N(0, σ 2 )
60 Posterior Inference Posterior averages are frequently important in problem domains posterior predictive distribution p(x i+1 x 1:i )= R p(x i+1 θ, x 1:i )p(θ x 1:i )dθ evidence (as seen) for model comparison, etc.
61 Relating the Posterior to the Posterior Predictive p(x i y 1:i ) = p(x i, x 1:i 1 y 1:i )dx 1:i 1 p(y i x i, x 1:i 1, y 1:i 1 )p(x i, x 1:i 1, y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x 1:i 1, y 1:i )p(x 1:i 1 y 1:i )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x 1:i 1 y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 p(y i x i )p(x i y 1:i 1 )
62 Relating the Posterior to the Posterior Predictive p(x i y 1:i ) = p(x i, x 1:i 1 y 1:i )dx 1:i 1 p(y i x i, x 1:i 1, y 1:i 1 )p(x i, x 1:i 1, y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x 1:i 1, y 1:i )p(x 1:i 1 y 1:i )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x 1:i 1 y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 p(y i x i )p(x i y 1:i 1 )
63 Relating the Posterior to the Posterior Predictive p(x i y 1:i ) = p(x i, x 1:i 1 y 1:i )dx 1:i 1 p(y i x i, x 1:i 1, y 1:i 1 )p(x i, x 1:i 1, y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x 1:i 1, y 1:i )p(x 1:i 1 y 1:i )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x 1:i 1 y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 p(y i x i )p(x i y 1:i 1 )
64 Relating the Posterior to the Posterior Predictive p(x i y 1:i ) = p(x i, x 1:i 1 y 1:i )dx 1:i 1 p(y i x i, x 1:i 1, y 1:i 1 )p(x i, x 1:i 1, y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x 1:i 1, y 1:i )p(x 1:i 1 y 1:i )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x 1:i 1 y 1:i 1 )dx 1:i 1 p(y i x i ) p(x i x i 1 )p(x i 1 y 1:i 1 )dx i 1 p(y i x i )p(x i y 1:i 1 )
Sequential Monte Carlo and Particle Filtering. Frank Wood Gatsby, November 2007
Sequential Monte Carlo and Particle Filtering Frank Wood Gatsby, November 2007 Importance Sampling Recall: Let s say that we want to compute some expectation (integral) E p [f] = p(x)f(x)dx and we remember
More informationMCMC and Gibbs Sampling. Kayhan Batmanghelich
MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationAnswers and expectations
Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E
More informationIntroduction to Bayesian inference
Introduction to Bayesian inference Thomas Alexander Brouwer University of Cambridge tab43@cam.ac.uk 17 November 2015 Probabilistic models Describe how data was generated using probability distributions
More informationIntroduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak
Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,
More informationAn Brief Overview of Particle Filtering
1 An Brief Overview of Particle Filtering Adam M. Johansen a.m.johansen@warwick.ac.uk www2.warwick.ac.uk/fac/sci/statistics/staff/academic/johansen/talks/ May 11th, 2010 Warwick University Centre for Systems
More informationVariational Scoring of Graphical Model Structures
Variational Scoring of Graphical Model Structures Matthew J. Beal Work with Zoubin Ghahramani & Carl Rasmussen, Toronto. 15th September 2003 Overview Bayesian model selection Approximations using Variational
More informationDavid Giles Bayesian Econometrics
David Giles Bayesian Econometrics 1. General Background 2. Constructing Prior Distributions 3. Properties of Bayes Estimators and Tests 4. Bayesian Analysis of the Multiple Regression Model 5. Bayesian
More informationCS281A/Stat241A Lecture 22
CS281A/Stat241A Lecture 22 p. 1/4 CS281A/Stat241A Lecture 22 Monte Carlo Methods Peter Bartlett CS281A/Stat241A Lecture 22 p. 2/4 Key ideas of this lecture Sampling in Bayesian methods: Predictive distribution
More informationMachine Learning Summer School
Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationL09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms
L09. PARTICLE FILTERING NA568 Mobile Robotics: Methods & Algorithms Particle Filters Different approach to state estimation Instead of parametric description of state (and uncertainty), use a set of state
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationPart 1: Expectation Propagation
Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud
More informationMonte Carlo Methods. Geoff Gordon February 9, 2006
Monte Carlo Methods Geoff Gordon ggordon@cs.cmu.edu February 9, 2006 Numerical integration problem 5 4 3 f(x,y) 2 1 1 0 0.5 0 X 0.5 1 1 0.8 0.6 0.4 Y 0.2 0 0.2 0.4 0.6 0.8 1 x X f(x)dx Used for: function
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin
More informationUnsupervised Learning
Unsupervised Learning Bayesian Model Comparison Zoubin Ghahramani zoubin@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc in Intelligent Systems, Dept Computer Science University College
More informationLecture 6: Graphical Models: Learning
Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)
More informationLearning Bayesian network : Given structure and completely observed data
Learning Bayesian network : Given structure and completely observed data Probabilistic Graphical Models Sharif University of Technology Spring 2017 Soleymani Learning problem Target: true distribution
More informationLecture 6: Bayesian Inference in SDE Models
Lecture 6: Bayesian Inference in SDE Models Bayesian Filtering and Smoothing Point of View Simo Särkkä Aalto University Simo Särkkä (Aalto) Lecture 6: Bayesian Inference in SDEs 1 / 45 Contents 1 SDEs
More informationBayesian Inference for Dirichlet-Multinomials
Bayesian Inference for Dirichlet-Multinomials Mark Johnson Macquarie University Sydney, Australia MLSS Summer School 1 / 50 Random variables and distributed according to notation A probability distribution
More informationLecture 13 : Variational Inference: Mean Field Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2017 Lecture 13 : Variational Inference: Mean Field Approximation Lecturer: Willie Neiswanger Scribes: Xupeng Tong, Minxing Liu 1 Problem Setup 1.1
More informationApproximate Bayesian inference
Approximate Bayesian inference Variational and Monte Carlo methods Christian A. Naesseth 1 Exchange rate data 0 20 40 60 80 100 120 Month Image data 2 1 Bayesian inference 2 Variational inference 3 Stochastic
More informationLecture 7 and 8: Markov Chain Monte Carlo
Lecture 7 and 8: Markov Chain Monte Carlo 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering University of Cambridge http://mlg.eng.cam.ac.uk/teaching/4f13/ Ghahramani
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationLinear Dynamical Systems
Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should
More informationBayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework
HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for
More information9 Forward-backward algorithm, sum-product on factor graphs
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous
More informationLecture 7: Optimal Smoothing
Department of Biomedical Engineering and Computational Science Aalto University March 17, 2011 Contents 1 What is Optimal Smoothing? 2 Bayesian Optimal Smoothing Equations 3 Rauch-Tung-Striebel Smoother
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering
ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:
More informationFundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner
Fundamentals CS 281A: Statistical Learning Theory Yangqing Jia Based on tutorial slides by Lester Mackey and Ariel Kleiner August, 2011 Outline 1 Probability 2 Statistics 3 Linear Algebra 4 Optimization
More informationAn introduction to Sequential Monte Carlo
An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods
More informationIntegrated Non-Factorized Variational Inference
Integrated Non-Factorized Variational Inference Shaobo Han, Xuejun Liao and Lawrence Carin Duke University February 27, 2014 S. Han et al. Integrated Non-Factorized Variational Inference February 27, 2014
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate
More informationTopic Modelling and Latent Dirichlet Allocation
Topic Modelling and Latent Dirichlet Allocation Stephen Clark (with thanks to Mark Gales for some of the slides) Lent 2013 Machine Learning for Language Processing: Lecture 7 MPhil in Advanced Computer
More informationLECTURE 15 Markov chain Monte Carlo
LECTURE 15 Markov chain Monte Carlo There are many settings when posterior computation is a challenge in that one does not have a closed form expression for the posterior distribution. Markov chain Monte
More informationCSC2535: Computation in Neural Networks Lecture 7: Variational Bayesian Learning & Model Selection
CSC2535: Computation in Neural Networks Lecture 7: Variational Bayesian Learning & Model Selection (non-examinable material) Matthew J. Beal February 27, 2004 www.variational-bayes.org Bayesian Model Selection
More informationGraphical Models and Kernel Methods
Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 7 Sequential Monte Carlo methods III 7 April 2017 Computer Intensive Methods (1) Plan of today s lecture
More informationWeakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0
More information27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling
10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel
More informationThe Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision
The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that
More informationRobotics. Mobile Robotics. Marc Toussaint U Stuttgart
Robotics Mobile Robotics State estimation, Bayes filter, odometry, particle filter, Kalman filter, SLAM, joint Bayes filter, EKF SLAM, particle SLAM, graph-based SLAM Marc Toussaint U Stuttgart DARPA Grand
More information28 : Approximate Inference - Distributed MCMC
10-708: Probabilistic Graphical Models, Spring 2015 28 : Approximate Inference - Distributed MCMC Lecturer: Avinava Dubey Scribes: Hakim Sidahmed, Aman Gupta 1 Introduction For many interesting problems,
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Contents in latter part Linear Dynamical Systems What is different from HMM? Kalman filter Its strength and limitation Particle Filter
More informationIntroduction to Particle Filters for Data Assimilation
Introduction to Particle Filters for Data Assimilation Mike Dowd Dept of Mathematics & Statistics (and Dept of Oceanography Dalhousie University, Halifax, Canada STATMOS Summer School in Data Assimila5on,
More informationParticle Filtering for Data-Driven Simulation and Optimization
Particle Filtering for Data-Driven Simulation and Optimization John R. Birge The University of Chicago Booth School of Business Includes joint work with Nicholas Polson. JRBirge INFORMS Phoenix, October
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September
More informationRecent Advances in Bayesian Inference Techniques
Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian
More informationExpectation Propagation Algorithm
Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,
More informationExercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters
Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for
More informationIntroduction to Gaussian Processes
Introduction to Gaussian Processes Iain Murray murray@cs.toronto.edu CSC255, Introduction to Machine Learning, Fall 28 Dept. Computer Science, University of Toronto The problem Learn scalar function of
More informationPMR Learning as Inference
Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning
More informationAuxiliary Particle Methods
Auxiliary Particle Methods Perspectives & Applications Adam M. Johansen 1 adam.johansen@bristol.ac.uk Oxford University Man Institute 29th May 2008 1 Collaborators include: Arnaud Doucet, Nick Whiteley
More information13: Variational inference II
10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational
More informationLecture 2: Priors and Conjugacy
Lecture 2: Priors and Conjugacy Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de May 6, 2014 Some nice courses Fred A. Hamprecht (Heidelberg U.) https://www.youtube.com/watch?v=j66rrnzzkow Michael I.
More informationIntroduc)on to Bayesian Methods
Introduc)on to Bayesian Methods Bayes Rule py x)px) = px! y) = px y)py) py x) = px y)py) px) px) =! px! y) = px y)py) y py x) = py x) =! y "! y px y)py) px y)py) px y)py) px y)py)dy Bayes Rule py x) =
More informationCalibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods
Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods Jonas Hallgren 1 1 Department of Mathematics KTH Royal Institute of Technology Stockholm, Sweden BFS 2012 June
More informationLecture : Probabilistic Machine Learning
Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning
More informationData Analysis and Uncertainty Part 2: Estimation
Data Analysis and Uncertainty Part 2: Estimation Instructor: Sargur N. University at Buffalo The State University of New York srihari@cedar.buffalo.edu 1 Topics in Estimation 1. Estimation 2. Desirable
More informationComputational Cognitive Science
Computational Cognitive Science Lecture 8: Frank Keller School of Informatics University of Edinburgh keller@inf.ed.ac.uk Based on slides by Sharon Goldwater October 14, 2016 Frank Keller Computational
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo (and Bayesian Mixture Models) David M. Blei Columbia University October 14, 2014 We have discussed probabilistic modeling, and have seen how the posterior distribution is the critical
More informationAdvanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering
Advanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering Axel Gandy Department of Mathematics Imperial College London http://www2.imperial.ac.uk/~agandy London
More informationBayesian Estimation with Sparse Grids
Bayesian Estimation with Sparse Grids Kenneth L. Judd and Thomas M. Mertens Institute on Computational Economics August 7, 27 / 48 Outline Introduction 2 Sparse grids Construction Integration with sparse
More informationan introduction to bayesian inference
with an application to network analysis http://jakehofman.com january 13, 2010 motivation would like models that: provide predictive and explanatory power are complex enough to describe observed phenomena
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More informationBayesian Machine Learning
Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 2: Bayesian Basics https://people.orie.cornell.edu/andrew/orie6741 Cornell University August 25, 2016 1 / 17 Canonical Machine Learning
More informationSMC 2 : an efficient algorithm for sequential analysis of state-space models
SMC 2 : an efficient algorithm for sequential analysis of state-space models N. CHOPIN 1, P.E. JACOB 2, & O. PAPASPILIOPOULOS 3 1 ENSAE-CREST 2 CREST & Université Paris Dauphine, 3 Universitat Pompeu Fabra
More informationLecture 8: Bayesian Estimation of Parameters in State Space Models
in State Space Models March 30, 2016 Contents 1 Bayesian estimation of parameters in state space models 2 Computational methods for parameter estimation 3 Practical parameter estimation in state space
More informationParameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn
Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation
More informationLearning of dynamical systems
Learning of dynamical systems Particle filters and Markov chain methods Thomas B. Schön and Fredrik Lindsten c Draft date August 23, 2017 2 Contents 1 Introduction 3 1.1 A few words for readers of the
More informationSampling Methods (11/30/04)
CS281A/Stat241A: Statistical Learning Theory Sampling Methods (11/30/04) Lecturer: Michael I. Jordan Scribe: Jaspal S. Sandhu 1 Gibbs Sampling Figure 1: Undirected and directed graphs, respectively, with
More informationComputation and Monte Carlo Techniques
Computation and Monte Carlo Techniques Up to now we have seen conjugate Bayesian analysis: posterior prior likelihood beta(a + x, b + n x) beta(a, b) binomial(x; n) θ a+x 1 (1 θ) b+n x 1 θ a 1 (1 θ) b
More informationSequential Monte Carlo Methods (for DSGE Models)
Sequential Monte Carlo Methods (for DSGE Models) Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 Some References These lectures use material from our joint work: Tempered
More informationLecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu
Lecture: Gaussian Process Regression STAT 6474 Instructor: Hongxiao Zhu Motivation Reference: Marc Deisenroth s tutorial on Robot Learning. 2 Fast Learning for Autonomous Robots with Gaussian Processes
More informationMCMC 2: Lecture 3 SIR models - more topics. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham
MCMC 2: Lecture 3 SIR models - more topics Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham Contents 1. What can be estimated? 2. Reparameterisation 3. Marginalisation
More informationLecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions
DD2431 Autumn, 2014 1 2 3 Classification with Probability Distributions Estimation Theory Classification in the last lecture we assumed we new: P(y) Prior P(x y) Lielihood x2 x features y {ω 1,..., ω K
More informationMarkov Chain Monte Carlo Methods for Stochastic Optimization
Markov Chain Monte Carlo Methods for Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U of Toronto, MIE,
More informationMarkov chain Monte Carlo Lecture 9
Markov chain Monte Carlo Lecture 9 David Sontag New York University Slides adapted from Eric Xing and Qirong Ho (CMU) Limitations of Monte Carlo Direct (unconditional) sampling Hard to get rare events
More informationParticle Filtering Approaches for Dynamic Stochastic Optimization
Particle Filtering Approaches for Dynamic Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge I-Sim Workshop,
More informationSensor Fusion: Particle Filter
Sensor Fusion: Particle Filter By: Gordana Stojceska stojcesk@in.tum.de Outline Motivation Applications Fundamentals Tracking People Advantages and disadvantages Summary June 05 JASS '05, St.Petersburg,
More informationVariational Inference via Stochastic Backpropagation
Variational Inference via Stochastic Backpropagation Kai Fan February 27, 2016 Preliminaries Stochastic Backpropagation Variational Auto-Encoding Related Work Summary Outline Preliminaries Stochastic Backpropagation
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationPractical Bayesian Optimization of Machine Learning. Learning Algorithms
Practical Bayesian Optimization of Machine Learning Algorithms CS 294 University of California, Berkeley Tuesday, April 20, 2016 Motivation Machine Learning Algorithms (MLA s) have hyperparameters that
More informationLecture 4: Probabilistic Learning
DD2431 Autumn, 2015 1 Maximum Likelihood Methods Maximum A Posteriori Methods Bayesian methods 2 Classification vs Clustering Heuristic Example: K-means Expectation Maximization 3 Maximum Likelihood Methods
More informationLinear Models A linear model is defined by the expression
Linear Models A linear model is defined by the expression x = F β + ɛ. where x = (x 1, x 2,..., x n ) is vector of size n usually known as the response vector. β = (β 1, β 2,..., β p ) is the transpose
More informationModeling and state estimation Examples State estimation Probabilities Bayes filter Particle filter. Modeling. CSC752 Autonomous Robotic Systems
Modeling CSC752 Autonomous Robotic Systems Ubbo Visser Department of Computer Science University of Miami February 21, 2017 Outline 1 Modeling and state estimation 2 Examples 3 State estimation 4 Probabilities
More informationMarkov Chain Monte Carlo Methods for Stochastic
Markov Chain Monte Carlo Methods for Stochastic Optimization i John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U Florida, Nov 2013
More informationSTA 4273H: Sta-s-cal Machine Learning
STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our
More informationIntroduction to Stochastic Gradient Markov Chain Monte Carlo Methods
Introduction to Stochastic Gradient Markov Chain Monte Carlo Methods Changyou Chen Department of Electrical and Computer Engineering, Duke University cc448@duke.edu Duke-Tsinghua Machine Learning Summer
More informationExpectation Propagation in Dynamical Systems
Expectation Propagation in Dynamical Systems Marc Peter Deisenroth Joint Work with Shakir Mohamed (UBC) August 10, 2012 Marc Deisenroth (TU Darmstadt) EP in Dynamical Systems 1 Motivation Figure : Complex
More informationBayesian GLMs and Metropolis-Hastings Algorithm
Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,
More information