Target Tracking and Classification using Collaborative Sensor Networks
|
|
- Caroline Robbins
- 5 years ago
- Views:
Transcription
1 Target Tracking and Classification using Collaborative Sensor Networks Xiaodong Wang Department of Electrical Engineering Columbia University
2 p.1/3 Talk Outline Background on distributed wireless sensor networks Single target tracking A Multiple target tracking and classification
3 Distributed Wireless Sensor Networks p.2/3
4 A sensor network is composed of a large number of sensor nodes, which are densely deployed either inside the phenomenon or very close to it. Random deployment Cooperative capabilities
5 Network Technology Computational Power Sensor Network Sensor Technology
6 MANET WSN High Mobility Peer to Peer model Low number of nodes Energy constraint not too critical No redundancy in data Connected to global network Commercial user end applications Low mobility (most cases!) Peer to peer or masterslave Large number of nodes Very severe energy constraint Some redundancy in data Local network Back end monitoring and data collection
7 Evolution of WSN Characteristic Generation1 Generation 2 Generation3 Timeline Size Large shoe box or bigger Pack of cards Weight Kilograms Grams Negligible Dust particle (MEMS) Node Architecture Separate sensing, processing and communication Integrated sensing, processing and communication Topology Pont to point, star Client server, peer to peer Integrated sensing, processing and communication Peer to peer Power supply lifetime Large batteries, hours to days AA batteries, days to weeks Solar, months to years Deployment Vehicle placed or air drop single sensors Hand emplaced Embedded, sprinkled left behind
8 p.3/3 Single Target Tracking via Collaborative Sensor Network Actual trajectory Estimated trajectory by OPT Estimated trajectory by PNN Estimated trajectory by NN Sensors Y (m) X (m)
9 p.4/3 Sensor network architecture Sensor nodes are scattered in a field. Nodes collect data and route data back to the sink via a multihop wireless medium. Each node has limited energy and only performs on-demand processing tasks upon a query. Otherwise it is a standby mode. Each node extracts relevant summary statistics from the raw data which are stored locally inside the node, and may be transmitted to other nodes upon request. To infer global information, the sensor network must facilitate efficient hierarchical information fusion.
10 p.5/3 Signal model Motion model: x t = {x t,y t,vt x,vy t } x t+ t = I 2 ti 2 0 I 2 x t + t2 2 I 2 ti 2 u t, Σ u = diag(σ 2 x, σ 2 y). Measurement model: Relative Distance Model: h(x t ) = K 10η log 10 ( (yt ȳ t ) 2 + (x t x t ) 2) Relative Angle Model: ( ) yt ȳ t h(x t ) = arctan x t x t
11 p.6/3 Cluttered noise model Measurements: Correct measurement if the target is detected Incorrect measurements from clutters. m f : number of false alarms. P(m f = k) = e λv (λv )m f m f!, k = 0, 1, 2,... V : volume of the observation area λ: number of false alarms per unit volume
12 p.7/3 Problem statement m t : number of measurements from target or false alarms at time t z t = {z m t } m t m=1 : set of measurements at time t Z t = {z j } t j=1 : set of measurements up to t Define events = {z m t is the target oriented measurements}, m = 1, 2,...,m t, θ m t θ 0 t = { none of the measurements at time t is the target oriented} P(θ m t m t ) = P D P G m t, m = 1,...,m t, 1 P D P G, m = 0. Goal: perform online estimation of p(x t Z t ) based on Z t at densely employed sensor nodes.
13 p.8/3 Sensor Selection Suppose at time t, sensor l is the leader node and holds the belief p(x t Z t ). For the new leader node at time t + 1, we select sensor s that achieves the maximum information gain of x t+1 conveyed by the new measurement z t+1,s I(x t+1,z t+1,s Z t ) = H(x t+1 Z t ) H(x t+1 Z t,z t+1,s ) In practice, the new observation z t+1,s is not available yet at time t.
14 p.9/3 Sensor Selection I(x t+1,z t+1,s Z t ) = H(x t+1 Z t ) H(x t+1 Z t,z t+1,s ) Need to estimate the measurement z t+1,s from the predicted belief at time t and use an expected likelihood function ˆp( z t+1,s x t+1 ) Then calculate the entropies using ˆp (x t+1 z t+1,s,z t ) = ˆp ( z t+1,s x t+1 )p(x t+1 Z t ).
15 p.10/3 SMC data fusion Let sensor l be the leader node at time t, which samples {X (j) t 1,w t 1} (j) N j=1 of p(x t 1 Z t 1 ). Let x t be the target state and z t = {z m t } m t m=1 be a set of validated measurements Denote X t = {X t 1,x t } and Z t = {Z t 1,z t }. We are interested in the on-line estimation of the new belief state p(x t Z t ). Use an auxiliary particle filter that resamples at the previous time step, based on some µ (j) t that characterize p(x t X (j) t 1).
16 p.11/3 SMC data fusion 1. For each i = 1,...,N, compute the point estimate of µ i t extended Kalman filter or one step prediction based on x i t For j = 1,...,N, do the following steps: (a) Sample an auxiliary integer variable from the set with probabilities proportional to by the g (i) t w (i) t 1 p(z t µ (i) t ); call the sampled index i (j). (b) Sample a value of the current state vector x (j) t xt 1 i(j). (c) Evaluate the corresponding weight. based on the
17 p.12/3 Entropy calculation Need to calculate the entropy of expected posterior distribution based on the discrete belief state {x (j) t,w (j) t } N j=1. Using the trial distribution q(x t+1 ) = p(x t+1 x (j) t,z t ), draw samples {X (j) t+1,w(j) t+1 )}N j=1 Then {X (j) t+1,w(j) t+1 )}N j=1 respect to p(x t+1 Z t ). with w(j) t+1 = w(j) t. is properly weighted with
18 Entropy calculation The expected posterior belief for sensor s can be represented by the discrete belief state {x (j) t+1, w(j) t+1,s }N j=1 w (j) t+1,s ˆp( z t+1,s x (j) t+1 ) = N Entropy calculation: = ˆp( z t+1,s x (j) t+1 )w(j) t+1 k=1 p (z t+1,s (x (k) t+1 ) x(j) t+1 )w(k) t+1 H(x t+1 Z t,z t+1,s ) = N j=1 w (j) t+1,s log w(j) t+1,s p.13/3
19 p.14/3 Kernel Smoothing Kernel density estimate based on samples: p(x t Z t ) = N w (j) t N ( x t m (j) t,h 2 V t ) j=1 with m (j) t = ax (j) t + (1 a) x t, a = 1 h 2. Instead of resampling from discrete samples, can resample from the continuous approximation, which can mitigate the degeneracy problem. Can collapse mixture components by replacing nearest pairs with a single averaged one reduce communication overhead between sensors.
20 p.15/3 Summary The initial leader node a does the following: 1. Draw initial samples {x (j) 0, w(j) 0 = 1} N j=1 information; 2. Update the belief state {x (j) 1, w(j) 1 }N j=1 measurement z 1 at leader node a; of the target from the prior based on the new 3. Compute the expected posterior belief state {x (j) 2, w(j) 2,i }N j=1 for each neighbor node i; 4. Compute the entropy of the expected posterior belief state {x (j) 2, w(j) 2,i }N j=1 for each neighbor node i, and determine the next best sensor, say b;
21 p.16/3 Summary 5 { Compute the kernel representation } of the belief state {m (j) 1, w(j) 1 }N j=1, h, V t and collapse the kernel smoothed belief state; 6 { Hand off the collapsed kernel representation { m (j) 1, w(j) 1 }K j=1, h, } V 1 to b. Node b takes the role of the leader, and it does the following: 1. Resample the belief state {x (j) 1, w(j) 1 = 1} N j=1 based on the the { collapsed kernel representation { m (j) 1, w(j) 1 }K j=1, h, } V 1 by the kernel-based reampling algorithm;
22 Summary 2 Update the belief state {x (j) 2, w(j) 2 = 1} N j=1 by the sensor fusion algorithm based on the new measurements z 2 and the belief state {x (j) 1, w(j) 1 = 1} N j=1 ; 3 Compute the expected posterior belief state {x (j) 3, w(j) 3,i }N j=1 for each neighbor node i with the weights w (j) 3,i ; 4 Compute the entropy on the expected posterior belief state {x (j) 3, w(j) 3,i }N j=1 for each neighbor node i, and determine the next best sensor, say c; 5 { Compute the kernel representation } of the belief state {m (j) 2, w(j) 2 }N j=1, h, V 2 and collapse it; 6 { Hand off the collapsed kernel representation { m (j) 2, w(j) 2 }K j=1, h, } V 2 to c; p.17/3
23 p.18/3 Simulation Results Localization s6 s6 bias Source Sensor Posterior mean estimate s5 s3 s5 50 Y (m) s1 s4 s2 s X (m) 1480 s s s X (m) X (m)
24 p.19/3 Simulation Results Tracking Actual trajectory Estimated trajectory by OPT Estimated trajectory by PNN Estimated trajectory by NN Sensors Y (m) X (m)
25 Multiple Target Tracking and Classification p.20/3
26 p.21/3 Multi-Target Tracking/Classification Densely scattered sensor nodes able to communicate with their neighbors and to process information. Multiple target tracking: sequential estimation the state of a possibly varying number of objects Classification: identification of those objects down to a given class motion model False detections: measurements arising from the clutter Unknown origin of the measurements complex data association problem
27 Multi-Target Tracking/Classification Sensor collaboration easier, localized tasks single target algorithm with (dis)appearing target Sensors Actual track 1 Actual track 2 Estimation, target 1 Selected sensors (1) Estimation, target 2 Selected sensors (2) Y(m) X(m) p.22/3
28 p.23/3 Leader-based Tracking Operations performed by a leader node [?] in multiple-target tracking and classification:
29 p.24/3 Target Dynamics The dynamic of the system is decomposed into x t,i = F γt,i (x t 1,i,u t,i ), i T t, γ t,i = γ t 1,i, i T t. T t, set of active targets x t,i, position and velocity of the i th target γ t,i, class of the i th target u t,i, independent white noise terms
30 Sensing Model The general model for the measurements is z m t = H t (x t,at (m),v t,m ), m {m a t (m ) 0}, z m t p c (z), m {m a t (m ) = 0}. z m t, m th measurement a t, data association vector H t, measurement function (e.g. power measurement) p c, clutter measurements distribution (e.g. uniform in the power range) The number of clutter measurements, m 0 t, typically arises from a poisson distribution. p.25/3
31 p.26/3 Measurements Likelihood Computation of the conditional distribution of the measurements enumeration of all possible data associations p(z t X t,r t ) = a t p(z t X t,r t,a t )p(a t r t,m t ). Having the dynamic of the system and the conditional distribution of the measurements we can now apply the SMC methodology to solve for the posterior distribution.
32 p.27/3 Single Target Tracking/Classification We compare several classes of dynamic models. 1 st approach: include the class in the state vector Problematic because of the fixed model for the class All particles eventually settle in one class can use a separate filter for each class We propose to modify the resampling scheme
33 Class-based Resampling Draw the number of particles for each class N γ according to { ˆP(γ Z 0:t )} γ Λ. If N γ < N Threshold, set N γ = N Threshold. Reduce the number of particles from the class with the most particles until γ Λ N γ = N. Draw N γ sample streams {X (J) 0:t } from {X (j) 0:t } j {j γ (j ) =γ} t weights {w (j) t } j {j γ (j ) t =γ}. with probability proportional to the Assign equal weight to each new sample within a class i.e. w (J) t = ˆP(γ (J) Z 0:t ) N. γ (J) p.28/3
34 p.29/3 Jump Markov Systems Varying number of targets can be dealt through Hypothesis testing Assuming a dynamic on the number of targets r t We consider a markovian evolution of r t π rt,r t 1 = p(r t r t 1 ) This model is referred as jump Markov Systems (JMS).
35 p.30/3 Choice of the Sampling Density Optimal density is intractable and its approximation are computationally heavy Main problem resides in the density for the number of targets we approximate only the optimal density for the number of targets p(r t X (j) t 1,r(j) t 1,Z t) p(z t X (j) t 1,r(j) t 1,r t)p(r t r (j) t 1 ) q(r t X (j) t 1,r(j) t 1,Z t) p(z t µ (j) t (r t ))p(r t r (j) t 1 ) µ (j) t,i = E[x t,i x (j) t 1,i,γ i]
36 p.31/3 Sensor Selection Scheme Information driven sensor selection s t, sensor chosen at time t s t+1 = arg max s ( Eps (Z t+1 Z 1:t ) [αυ utility (s) + (1 + α)υ cost (s)] ) E[Υ utility (s)] = I s (x t+1,1 ; z t+1 Z 1:t ) Approximation using Monte Carlo Integration from the available samples
37 p.32/3 Simulation Scenario We consider the tracking of two crossing targets from 2 different classes The first is initially tracked The second is initially unknown Probability of detection: 0.95 Average clutter measurements/time step: 1 Field: m 2 covered by 300 randomly scattered sensor nodes Sensing range: [4m, 50m]
38 Tracking Results Actual and estimated trajectories with an unknown target Sensors Actual track 1 Actual track 2 Estimation, target 1 Selected sensors (1) Estimation, target 2 Selected sensors (2) Y(m) X(m) p.33/3
39 p.34/3 Target Detection Results Probabilities of having two targets in the field: (left) first leader node; (right) newly generated second leader node. Probability of having two targets Time Probability of having two targets Generation of the second leader node Time
40 p.35/3 Classification Results Probabilities that the first target of the leader node is from the first class: (left) first leader node; (right) newly generated second leader node. 1 1 Probability of being in class Time Probability of being in class Generation of the second leader node Time
41 p.36/3 Conclusions SMC provides a unique capability for supporting in-network distributed signal processing in non-linear and/or non-gaussian environments. Developed leader-based tracking multi-target tracking and classification algorithm based on SMC filtering. Developed a class-based resampling procedure to avoid the loss of plausible classification hypothesis during the early stage of tracking. Developed an SMC implementation of an optimal sensor selection scheme based on expected information gain.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Contents in latter part Linear Dynamical Systems What is different from HMM? Kalman filter Its strength and limitation Particle Filter
More informationContent.
Content Fundamentals of Bayesian Techniques (E. Sucar) Bayesian Filters (O. Aycard) Definition & interests Implementations Hidden Markov models Discrete Bayesian Filters or Markov localization Kalman filters
More informationL09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms
L09. PARTICLE FILTERING NA568 Mobile Robotics: Methods & Algorithms Particle Filters Different approach to state estimation Instead of parametric description of state (and uncertainty), use a set of state
More informationDistributed estimation in sensor networks
in sensor networks A. Benavoli Dpt. di Sistemi e Informatica Università di Firenze, Italy. e-mail: benavoli@dsi.unifi.it Outline 1 An introduction to 2 3 An introduction to An introduction to In recent
More informationApproximate Inference Part 1 of 2
Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ Bayesian paradigm Consistent use of probability theory
More informationApproximate Inference Part 1 of 2
Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ 1 Bayesian paradigm Consistent use of probability theory
More informationSequential Monte Carlo and Particle Filtering. Frank Wood Gatsby, November 2007
Sequential Monte Carlo and Particle Filtering Frank Wood Gatsby, November 2007 Importance Sampling Recall: Let s say that we want to compute some expectation (integral) E p [f] = p(x)f(x)dx and we remember
More informationSensor Fusion: Particle Filter
Sensor Fusion: Particle Filter By: Gordana Stojceska stojcesk@in.tum.de Outline Motivation Applications Fundamentals Tracking People Advantages and disadvantages Summary June 05 JASS '05, St.Petersburg,
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 7 Sequential Monte Carlo methods III 7 April 2017 Computer Intensive Methods (1) Plan of today s lecture
More informationParticle Filters. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics
Particle Filters Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Motivation For continuous spaces: often no analytical formulas for Bayes filter updates
More informationRao-Blackwellized Particle Filter for Multiple Target Tracking
Rao-Blackwellized Particle Filter for Multiple Target Tracking Simo Särkkä, Aki Vehtari, Jouko Lampinen Helsinki University of Technology, Finland Abstract In this article we propose a new Rao-Blackwellized
More informationSensor Tasking and Control
Sensor Tasking and Control Sensing Networking Leonidas Guibas Stanford University Computation CS428 Sensor systems are about sensing, after all... System State Continuous and Discrete Variables The quantities
More informationExpectation propagation for signal detection in flat-fading channels
Expectation propagation for signal detection in flat-fading channels Yuan Qi MIT Media Lab Cambridge, MA, 02139 USA yuanqi@media.mit.edu Thomas Minka CMU Statistics Department Pittsburgh, PA 15213 USA
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationThe Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision
The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that
More informationApproximate Inference
Approximate Inference Simulation has a name: sampling Sampling is a hot topic in machine learning, and it s really simple Basic idea: Draw N samples from a sampling distribution S Compute an approximate
More informationMulti-Robotic Systems
CHAPTER 9 Multi-Robotic Systems The topic of multi-robotic systems is quite popular now. It is believed that such systems can have the following benefits: Improved performance ( winning by numbers ) Distributed
More informationBayesian Networks BY: MOHAMAD ALSABBAGH
Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional
More informationAUTOMOTIVE ENVIRONMENT SENSORS
AUTOMOTIVE ENVIRONMENT SENSORS Lecture 5. Localization BME KÖZLEKEDÉSMÉRNÖKI ÉS JÁRMŰMÉRNÖKI KAR 32708-2/2017/INTFIN SZÁMÚ EMMI ÁLTAL TÁMOGATOTT TANANYAG Related concepts Concepts related to vehicles moving
More informationIntroduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization
Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Kai Arras 1 Motivation Recall: Discrete filter Discretize the
More informationPROBABILISTIC REASONING OVER TIME
PROBABILISTIC REASONING OVER TIME In which we try to interpret the present, understand the past, and perhaps predict the future, even when very little is crystal clear. Outline Time and uncertainty Inference:
More informationArtificial Intelligence
Artificial Intelligence Roman Barták Department of Theoretical Computer Science and Mathematical Logic Summary of last lecture We know how to do probabilistic reasoning over time transition model P(X t
More informationMultiple Bits Distributed Moving Horizon State Estimation for Wireless Sensor Networks. Ji an Luo
Multiple Bits Distributed Moving Horizon State Estimation for Wireless Sensor Networks Ji an Luo 2008.6.6 Outline Background Problem Statement Main Results Simulation Study Conclusion Background Wireless
More informationA Probabilistic Framework for solving Inverse Problems. Lambros S. Katafygiotis, Ph.D.
A Probabilistic Framework for solving Inverse Problems Lambros S. Katafygiotis, Ph.D. OUTLINE Introduction to basic concepts of Bayesian Statistics Inverse Problems in Civil Engineering Probabilistic Model
More informationLecture 13 : Variational Inference: Mean Field Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2017 Lecture 13 : Variational Inference: Mean Field Approximation Lecturer: Willie Neiswanger Scribes: Xupeng Tong, Minxing Liu 1 Problem Setup 1.1
More informationKernel adaptive Sequential Monte Carlo
Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline
More informationAuxiliary Particle Methods
Auxiliary Particle Methods Perspectives & Applications Adam M. Johansen 1 adam.johansen@bristol.ac.uk Oxford University Man Institute 29th May 2008 1 Collaborators include: Arnaud Doucet, Nick Whiteley
More informationAnswers and expectations
Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E
More informationSEQUENTIAL MONTE CARLO METHODS WITH APPLICATIONS TO COMMUNICATION CHANNELS. A Thesis SIRISH BODDIKURAPATI
SEQUENTIAL MONTE CARLO METHODS WITH APPLICATIONS TO COMMUNICATION CHANNELS A Thesis by SIRISH BODDIKURAPATI Submitted to the Office of Graduate Studies of Texas A&M University in partial fulfillment of
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationMCMC and Gibbs Sampling. Kayhan Batmanghelich
MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 18: HMMs and Particle Filtering 4/4/2011 Pieter Abbeel --- UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew Moore
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationEVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER
EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER Zhen Zhen 1, Jun Young Lee 2, and Abdus Saboor 3 1 Mingde College, Guizhou University, China zhenz2000@21cn.com 2 Department
More informationPart 1: Expectation Propagation
Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud
More informationModeling and state estimation Examples State estimation Probabilities Bayes filter Particle filter. Modeling. CSC752 Autonomous Robotic Systems
Modeling CSC752 Autonomous Robotic Systems Ubbo Visser Department of Computer Science University of Miami February 21, 2017 Outline 1 Modeling and state estimation 2 Examples 3 State estimation 4 Probabilities
More informationNotes on Machine Learning for and
Notes on Machine Learning for 16.410 and 16.413 (Notes adapted from Tom Mitchell and Andrew Moore.) Choosing Hypotheses Generally want the most probable hypothesis given the training data Maximum a posteriori
More information27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling
10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel
More informationGraphical Models and Kernel Methods
Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.
More informationOutline. 1 Introduction. 2 Problem and solution. 3 Bayesian tracking model of group debris. 4 Simulation results. 5 Conclusions
Outline 1 Introduction 2 Problem and solution 3 Bayesian tracking model of group debris 4 Simulation results 5 Conclusions Problem The limited capability of the radar can t always satisfy the detection
More informationMachine Learning for Data Science (CS4786) Lecture 24
Machine Learning for Data Science (CS4786) Lecture 24 Graphical Models: Approximate Inference Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ BELIEF PROPAGATION OR MESSAGE PASSING Each
More informationProbabilistic Fundamentals in Robotics. DAUIN Politecnico di Torino July 2010
Probabilistic Fundamentals in Robotics Probabilistic Models of Mobile Robots Robot localization Basilio Bona DAUIN Politecnico di Torino July 2010 Course Outline Basic mathematical framework Probabilistic
More informationBayesian Methods in Positioning Applications
Bayesian Methods in Positioning Applications Vedran Dizdarević v.dizdarevic@tugraz.at Graz University of Technology, Austria 24. May 2006 Bayesian Methods in Positioning Applications p.1/21 Outline Problem
More informationMobile Robot Localization
Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations
More informationBayesian Networks Inference with Probabilistic Graphical Models
4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning
More informationProbabilistic Robotics
University of Rome La Sapienza Master in Artificial Intelligence and Robotics Probabilistic Robotics Prof. Giorgio Grisetti Course web site: http://www.dis.uniroma1.it/~grisetti/teaching/probabilistic_ro
More informationChapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang
Chapter 4 Dynamic Bayesian Networks 2016 Fall Jin Gu, Michael Zhang Reviews: BN Representation Basic steps for BN representations Define variables Define the preliminary relations between variables Check
More informationAdaptive Multi-Modal Sensing of General Concealed Targets
Adaptive Multi-Modal Sensing of General Concealed argets Lawrence Carin Balaji Krishnapuram, David Williams, Xuejun Liao and Ya Xue Department of Electrical & Computer Engineering Duke University Durham,
More informationProbabilistic Fundamentals in Robotics
Probabilistic Fundamentals in Robotics Probabilistic Models of Mobile Robots Robot localization Basilio Bona DAUIN Politecnico di Torino June 2011 Course Outline Basic mathematical framework Probabilistic
More informationComputer Vision Group Prof. Daniel Cremers. 14. Sampling Methods
Prof. Daniel Cremers 14. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationDynamic Bandwidth Allocation for Target Tracking. Wireless Sensor Networks.
4th International Conference on Information Fusion Chicago, Illinois, USA, July 5-8, Dynamic Bandwidth Allocation for Target Tracking in Wireless Sensor Networks Engin Masazade, Ruixin Niu, and Pramod
More information15-780: Grad AI Lecture 19: Graphical models, Monte Carlo methods. Geoff Gordon (this lecture) Tuomas Sandholm TAs Erik Zawadzki, Abe Othman
15-780: Grad AI Lecture 19: Graphical models, Monte Carlo methods Geoff Gordon (this lecture) Tuomas Sandholm TAs Erik Zawadzki, Abe Othman Admin Reminder: midterm March 29 Reminder: project milestone
More informationAdaptive Particle Filter for Nonparametric Estimation with Measurement Uncertainty in Wireless Sensor Networks
Article Adaptive Particle Filter for Nonparametric Estimation with Measurement Uncertainty in Wireless Sensor Networks Xiaofan Li 1,2,, Yubin Zhao 3, *, Sha Zhang 1,2, and Xiaopeng Fan 3 1 The State Monitoring
More informationA Unifying Framework for Multi-Target Tracking and Existence
A Unifying Framework for Multi-Target Tracking and Existence Jaco Vermaak Simon Maskell, Mark Briers Cambridge University Engineering Department QinetiQ Ltd., Malvern Technology Centre Cambridge, U.K.
More informationParticle Filtering a brief introductory tutorial. Frank Wood Gatsby, August 2007
Particle Filtering a brief introductory tutorial Frank Wood Gatsby, August 2007 Problem: Target Tracking A ballistic projectile has been launched in our direction and may or may not land near enough to
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationA Comparison of the EKF, SPKF, and the Bayes Filter for Landmark-Based Localization
A Comparison of the EKF, SPKF, and the Bayes Filter for Landmark-Based Localization and Timothy D. Barfoot CRV 2 Outline Background Objective Experimental Setup Results Discussion Conclusion 2 Outline
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 13: Learning in Gaussian Graphical Models, Non-Gaussian Inference, Monte Carlo Methods Some figures
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering
ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:
More informationProbabilistic Graphical Models Lecture 17: Markov chain Monte Carlo
Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,
More informationCSEP 573: Artificial Intelligence
CSEP 573: Artificial Intelligence Hidden Markov Models Luke Zettlemoyer Many slides over the course adapted from either Dan Klein, Stuart Russell, Andrew Moore, Ali Farhadi, or Dan Weld 1 Outline Probabilistic
More informationLecture 8: Bayesian Estimation of Parameters in State Space Models
in State Space Models March 30, 2016 Contents 1 Bayesian estimation of parameters in state space models 2 Computational methods for parameter estimation 3 Practical parameter estimation in state space
More informationDynamic System Identification using HDMR-Bayesian Technique
Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in
More informationHidden Markov Models. Vibhav Gogate The University of Texas at Dallas
Hidden Markov Models Vibhav Gogate The University of Texas at Dallas Intro to AI (CS 4365) Many slides over the course adapted from either Dan Klein, Luke Zettlemoyer, Stuart Russell or Andrew Moore 1
More informationSLAM Techniques and Algorithms. Jack Collier. Canada. Recherche et développement pour la défense Canada. Defence Research and Development Canada
SLAM Techniques and Algorithms Jack Collier Defence Research and Development Canada Recherche et développement pour la défense Canada Canada Goals What will we learn Gain an appreciation for what SLAM
More informationKalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein
Kalman filtering and friends: Inference in time series models Herke van Hoof slides mostly by Michael Rubinstein Problem overview Goal Estimate most probable state at time k using measurement up to time
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 12 Dynamical Models CS/CNS/EE 155 Andreas Krause Homework 3 out tonight Start early!! Announcements Project milestones due today Please email to TAs 2 Parameter learning
More informationState-Space Methods for Inferring Spike Trains from Calcium Imaging
State-Space Methods for Inferring Spike Trains from Calcium Imaging Joshua Vogelstein Johns Hopkins April 23, 2009 Joshua Vogelstein (Johns Hopkins) State-Space Calcium Imaging April 23, 2009 1 / 78 Outline
More informationRobust Monte Carlo Methods for Sequential Planning and Decision Making
Robust Monte Carlo Methods for Sequential Planning and Decision Making Sue Zheng, Jason Pacheco, & John Fisher Sensing, Learning, & Inference Group Computer Science & Artificial Intelligence Laboratory
More informationCollaborative Target Detection in Wireless Sensor Networks with Reactive Mobility
1 / 24 Collaborative Target Detection in Wireless Sensor Networks with Reactive Mobility Rui Tan 1 Guoliang Xing 1 Jianping Wang 1 Hing Cheung So 2 1 Department of Computer Science City University of Hong
More informationBearings-Only Tracking in Modified Polar Coordinate System : Initialization of the Particle Filter and Posterior Cramér-Rao Bound
Bearings-Only Tracking in Modified Polar Coordinate System : Initialization of the Particle Filter and Posterior Cramér-Rao Bound Thomas Brehard (IRISA/CNRS), Jean-Pierre Le Cadre (IRISA/CNRS). Journée
More information2D Image Processing (Extended) Kalman and particle filter
2D Image Processing (Extended) Kalman and particle filter Prof. Didier Stricker Dr. Gabriele Bleser Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz
More informationExercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters
Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for
More informationAn Brief Overview of Particle Filtering
1 An Brief Overview of Particle Filtering Adam M. Johansen a.m.johansen@warwick.ac.uk www2.warwick.ac.uk/fac/sci/statistics/staff/academic/johansen/talks/ May 11th, 2010 Warwick University Centre for Systems
More informationAdvanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering
Advanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering Axel Gandy Department of Mathematics Imperial College London http://www2.imperial.ac.uk/~agandy London
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods
Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationChapter 16. Structured Probabilistic Models for Deep Learning
Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin
More informationResults: MCMC Dancers, q=10, n=500
Motivation Sampling Methods for Bayesian Inference How to track many INTERACTING targets? A Tutorial Frank Dellaert Results: MCMC Dancers, q=10, n=500 1 Probabilistic Topological Maps Results Real-Time
More informationAnnouncements. Proposals graded
Announcements Proposals graded Kevin Jamieson 2018 1 Bayesian Methods Machine Learning CSE546 Kevin Jamieson University of Washington November 1, 2018 2018 Kevin Jamieson 2 MLE Recap - coin flips Data:
More informationPartially Observable Markov Decision Processes (POMDPs)
Partially Observable Markov Decision Processes (POMDPs) Sachin Patil Guest Lecture: CS287 Advanced Robotics Slides adapted from Pieter Abbeel, Alex Lee Outline Introduction to POMDPs Locally Optimal Solutions
More informationRAO-BLACKWELLIZED PARTICLE FILTER FOR MARKOV MODULATED NONLINEARDYNAMIC SYSTEMS
RAO-BLACKWELLIZED PARTICLE FILTER FOR MARKOV MODULATED NONLINEARDYNAMIC SYSTEMS Saiat Saha and Gustaf Hendeby Linöping University Post Print N.B.: When citing this wor, cite the original article. 2014
More informationRobotics. Mobile Robotics. Marc Toussaint U Stuttgart
Robotics Mobile Robotics State estimation, Bayes filter, odometry, particle filter, Kalman filter, SLAM, joint Bayes filter, EKF SLAM, particle SLAM, graph-based SLAM Marc Toussaint U Stuttgart DARPA Grand
More informationProbabilistic Reasoning in Deep Learning
Probabilistic Reasoning in Deep Learning Dr Konstantina Palla, PhD palla@stats.ox.ac.uk September 2017 Deep Learning Indaba, Johannesburgh Konstantina Palla 1 / 39 OVERVIEW OF THE TALK Basics of Bayesian
More informationLecture 6: Bayesian Inference in SDE Models
Lecture 6: Bayesian Inference in SDE Models Bayesian Filtering and Smoothing Point of View Simo Särkkä Aalto University Simo Särkkä (Aalto) Lecture 6: Bayesian Inference in SDEs 1 / 45 Contents 1 SDEs
More informationAn introduction to Sequential Monte Carlo
An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods
More informationSequential Monte Carlo Samplers for Applications in High Dimensions
Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex
More informationCS242: Probabilistic Graphical Models Lecture 4B: Learning Tree-Structured and Directed Graphs
CS242: Probabilistic Graphical Models Lecture 4B: Learning Tree-Structured and Directed Graphs Professor Erik Sudderth Brown University Computer Science October 6, 2016 Some figures and materials courtesy
More informationUWB Geolocation Techniques for IEEE a Personal Area Networks
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com UWB Geolocation Techniques for IEEE 802.15.4a Personal Area Networks Sinan Gezici Zafer Sahinoglu TR-2004-110 August 2004 Abstract A UWB positioning
More information13: Variational inference II
10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational
More informationL11. EKF SLAM: PART I. NA568 Mobile Robotics: Methods & Algorithms
L11. EKF SLAM: PART I NA568 Mobile Robotics: Methods & Algorithms Today s Topic EKF Feature-Based SLAM State Representation Process / Observation Models Landmark Initialization Robot-Landmark Correlation
More informationParameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1
Parameter Estimation William H. Jefferys University of Texas at Austin bill@bayesrules.net Parameter Estimation 7/26/05 1 Elements of Inference Inference problems contain two indispensable elements: Data
More informationSequential Bayesian Updating
BS2 Statistical Inference, Lectures 14 and 15, Hilary Term 2009 May 28, 2009 We consider data arriving sequentially X 1,..., X n,... and wish to update inference on an unknown parameter θ online. In a
More informationParticle lter for mobile robot tracking and localisation
Particle lter for mobile robot tracking and localisation Tinne De Laet K.U.Leuven, Dept. Werktuigkunde 19 oktober 2005 Particle lter 1 Overview Goal Introduction Particle lter Simulations Particle lter
More informationImplementation of Particle Filter-based Target Tracking
of -based V. Rajbabu rajbabu@ee.iitb.ac.in VLSI Group Seminar Dept. of Electrical Engineering IIT Bombay November 15, 2007 Outline Introduction 1 Introduction 2 3 4 5 2 / 55 Outline Introduction 1 Introduction
More informationComputational Biology Lecture #3: Probability and Statistics. Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept
Computational Biology Lecture #3: Probability and Statistics Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept 26 2005 L2-1 Basic Probabilities L2-2 1 Random Variables L2-3 Examples
More informationTemporal probability models. Chapter 15
Temporal probability models Chapter 15 Outline Time and uncertainty Inference: filtering, prediction, smoothing Hidden Markov models Kalman filters (a brief mention) Dynamic Bayesian networks Particle
More informationSequential Monte Carlo Methods (for DSGE Models)
Sequential Monte Carlo Methods (for DSGE Models) Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 Some References These lectures use material from our joint work: Tempered
More information