CONDENSATION Conditional Density Propagation for Visual Tracking

Similar documents
2D Image Processing (Extended) Kalman and particle filter

2D Image Processing. Bayes filter implementation: Kalman filter

2D Image Processing. Bayes filter implementation: Kalman filter

CSC487/2503: Foundations of Computer Vision. Visual Tracking. David Fleet

Partially Observable Markov Decision Processes (POMDPs)

Probabilistic Graphical Models

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

Particle Filters. Outline

Human Pose Tracking I: Basics. David Fleet University of Toronto

Introduction to Machine Learning

Markov chain Monte Carlo methods for visual tracking

Statistical Techniques in Robotics (16-831, F12) Lecture#17 (Wednesday October 31) Kalman Filters. Lecturer: Drew Bagnell Scribe:Greydon Foil 1

Tracking using CONDENSATION: Conditional Density Propagation

Application of new Monte Carlo method for inversion of prestack seismic data. Yang Xue Advisor: Dr. Mrinal K. Sen

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

Bayes Filter Reminder. Kalman Filter Localization. Properties of Gaussians. Gaussians. Prediction. Correction. σ 2. Univariate. 1 2πσ e.

Parallel Particle Filter in Julia

EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER

Density Propagation for Continuous Temporal Chains Generative and Discriminative Models

Nonlinear and/or Non-normal Filtering. Jesús Fernández-Villaverde University of Pennsylvania

Dynamic Multipath Estimation by Sequential Monte Carlo Methods

ROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino

EKF and SLAM. McGill COMP 765 Sept 18 th, 2017

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization

Mobile Robot Localization

Supplementary Note on Bayesian analysis

Dynamic System Identification using HDMR-Bayesian Technique

Rao-Blackwellized Particle Filter for Multiple Target Tracking

(Extended) Kalman Filter

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

The Kalman Filter (part 1) Definition. Rudolf Emil Kalman. Why do we need a filter? Definition. HCI/ComS 575X: Computational Perception.

AUTOMOTIVE ENVIRONMENT SENSORS

Tracking. Readings: Chapter 17 of Forsyth and Ponce. Matlab Tutorials: motiontutorial.m, trackingtutorial.m

Mobile Robot Localization

Face Recognition from Video: A CONDENSATION Approach

Fundamentals of Data Assimilation

Sensor Fusion: Particle Filter

The Kalman Filter ImPr Talk

F denotes cumulative density. denotes probability density function; (.)

Why do we care? Measurements. Handling uncertainty over time: predicting, estimating, recognizing, learning. Dealing with time

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Probabilistic Fundamentals in Robotics. DAUIN Politecnico di Torino July 2010

Robotics 2 Target Tracking. Kai Arras, Cyrill Stachniss, Maren Bennewitz, Wolfram Burgard

Clustering with k-means and Gaussian mixture distributions

CIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions

Kalman Filter Computer Vision (Kris Kitani) Carnegie Mellon University

A new Hierarchical Bayes approach to ensemble-variational data assimilation

Optical Flow, Motion Segmentation, Feature Tracking

Benefits of a Geosynchronous Orbit (GEO) Observation Point for Orbit Determination

Robotics 2 Target Tracking. Giorgio Grisetti, Cyrill Stachniss, Kai Arras, Wolfram Burgard

Multimodal Nested Sampling

An introduction to particle filters

In Situ Evaluation of Tracking Algorithms Using Time Reversed Chains

Miscellaneous. Regarding reading materials. Again, ask questions (if you have) and ask them earlier

Why do we care? Examples. Bayes Rule. What room am I in? Handling uncertainty over time: predicting, estimating, recognizing, learning

Learning Static Parameters in Stochastic Processes

CS 4495 Computer Vision Principle Component Analysis

STA 4273H: Statistical Machine Learning

Bayesian Methods / G.D. Hager S. Leonard

Autonomous Navigation for Flying Robots

Robot Localisation. Henrik I. Christensen. January 12, 2007

L09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms

A Study of Covariances within Basic and Extended Kalman Filters

Autonomous Mobile Robot Design

Bayesian Regression Linear and Logistic Regression

Blind Equalization via Particle Filtering

Relative Merits of 4D-Var and Ensemble Kalman Filter

Efficient Monitoring for Planetary Rovers

ERROR COVARIANCE ESTIMATION IN OBJECT TRACKING SCENARIOS USING KALMAN FILTER

Least Squares Estimation Namrata Vaswani,

Particle Filters. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics

Robotics. Mobile Robotics. Marc Toussaint U Stuttgart

Data assimilation in high dimensions

ABSTRACT INTRODUCTION

Robust Ensemble Filtering With Improved Storm Surge Forecasting


Non-Gaussian likelihoods for Gaussian Processes

Tracking Human Heads Based on Interaction between Hypotheses with Certainty

Probabilistic Fundamentals in Robotics

Ensemble Data Assimilation and Uncertainty Quantification

Results: MCMC Dancers, q=10, n=500

Machine Learning Techniques for Computer Vision

Karl-Rudolf Koch Introduction to Bayesian Statistics Second Edition

Markov Models. CS 188: Artificial Intelligence Fall Example. Mini-Forward Algorithm. Stationary Distributions.

COS Lecture 16 Autonomous Robot Navigation

The Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance

arxiv: v1 [stat.co] 23 Apr 2018

Gaussian Processes for Sequential Prediction

Probabilistic Fundamentals in Robotics. DAUIN Politecnico di Torino July 2010

Forecasting Wind Ramps

ISOMAP TRACKING WITH PARTICLE FILTER

Basics of Uncertainty Analysis

An introduction to Bayesian statistics and model calibration and a host of related topics

New Fast Kalman filter method

Lagrangian Data Assimilation and Manifold Detection for a Point-Vortex Model. David Darmon, AMSC Kayo Ide, AOSC, IPST, CSCAMM, ESSIC

Multi-Target Particle Filtering for the Probability Hypothesis Density

Hidden Markov Models. Hal Daumé III. Computer Science University of Maryland CS 421: Introduction to Artificial Intelligence 19 Apr 2012

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering

Particle Filters for Real-Time Fault Detection in Planetary Rovers

DATA ASSIMILATION FOR FLOOD FORECASTING

Transcription:

CONDENSATION Conditional Density Propagation for Visual Tracking Michael Isard and Andrew Blake Presented by Neil Alldrin Department of Computer Science & Engineering University of California, San Diego La Jolla, CA 92093, USA CONDENSATION p./40

Outline The problem / motivation Previous methods - Kalman filters CONDENSATION Modeling shape and motion Factored sampling The CONDENSATION algorithm Benefits / limitations Demonstration CONDENSATION p.2/40

The Goal What are we trying to do? Track an object throughout a video sequence in real-time Object can move and change shape CONDENSATION p.3/40

Motivation Tracking has many potential applications Surveillance Missile guidance systems Virtual mouse pointer Tracking is also useful for more advanced techniques such as behavior recognition CONDENSATION p.4/40

Difficulties Some difficulties specific to tracking include Object representation Background clutter Computational efficiency Initialization Occlusion CONDENSATION p.5/40

Existing Methods Tracking via repeated recognition (Sullivan & Carlson; Mori & Malik) Articulated Tracking (Breglar; Blark) Kalman Filters (Kalman - 960) CONDENSATION p.6/40

Kalman Filters Basic Idea Model the object being tracked At each time-frame, Prediction - Predict where object should be Measurement - Observe where object went Assimilation - Update object model by combining the two CONDENSATION p.7/40

Modeling shape and motion System Model Object modelled by a state vector, equations, the system model The state is a time-dependent vector variables, and a set of of system The system model is a vector equation describing the evolution of the state in time CONDENSATION p.8/40

Modeling shape and motion Assumptions The state does not change much between consecutive time instants Linear system model is the state transition matrix at time is a random vector modelling additive system noise CONDENSATION p.9/40

Modeling shape and motion Example Want to model a basketball Parameterize as a circle, CONDENSATION p.0/40

Modeling shape and motion is used to predict the next Example The state transition matrix object state......!!! CONDENSATION p./40

$ Modeling shape and motion Measurement Model At each time step a measurement, is taken, of the state vector Assume a linear relationship between measurements and the system state " " # # %$ is the measurement matrix is a random vector modelling additive noise CONDENSATION p.2/40

Modeling shape and motion Uncertainty Problem : The system model and feature measurements are imprecise Solution : Quantify the uncertainty with a probability density function, specifically a Gaussian Define covariance matrix vector & corresponding to state CONDENSATION p.3/40

The Kalman Filter Algorithm Want to estimate the system s state,, at time using the system model prediction from time and the measurements,, taken at time Three steps Prediction Measurement Assimilation " ' CONDENSATION p.4/40

' ' (& ' ' $ The Algorithm Symbols is the state estimate predicted before obtaining the measurement ( ' " is the state estimate computed after integrating the measurement with the prediction and ' & " ( are the state covariance matrices for respectively ) ( and is the gain matrix, specifying the relative importance of the prediction and the measurement * ( and are the covariance matrices of the noise processes and + ' CONDENSATION p.5/40

' (& The Algorithm Prediction Based on object state and system model from time, predict the object state at time ' ( * & CONDENSATION p.6/40

The Algorithm Measurement Extract features, ", from the image at time " is assumed to be a linear transformation of the object state, " # CONDENSATION p.7/40

' # (& #, The Algorithm Assimilation Combine the predicted state with the measured state to generate the new object state ( ' # " ) ( ' is the difference between the measured features and the predicted features " ) ( ' # (& # + Update the covariance matrix, & (& # ) & CONDENSATION p.8/40

The Algorithm Probability density propagation CONDENSATION p.9/40

() (2) ( & (3) # ( & # ( & (4) (5) ( & ) CONDENSATION p.20/40 The Algorithm Recap : The Kalman filter equations ' ( ' * & + # ) ' # " ) ' ' + ) ), ), &

Limitations Unimodal, Gaussian probability distribution Object state often non-gaussian For example, what if the object is likely to be in one of two locations, but not inbetween? Sensitive to background clutter CONDENSATION p.2/40

CONDENSATION Motivation Overcome limitations of Kalman filters; Specifically, allow multimodal probability distributions Increase robustness to background clutter CONDENSATION p.22/40

CONDENSATION Overview The primary difference between Kalman filters and the CONDENSATION algorithm is how the state probability density is represented. Kalman filters use a Gaussian CONDENSATION uses factored sampling to approximate arbitrary pdf s Other names for this algorithm Particle filters Monte-Carlo methods CONDENSATION p.23/40

CONDENSATION Outline Modelling shape and motion Factored sampling The algorithm Benefits Limitations CONDENSATION p.24/40

/.!!! /.!!! 2 Modelling Shape and Motion Symbols " - 0 is the object state vector at time is the vector of measured features at time " " " is the state history is the history of measured features is the prior probability density for object is the observation density CONDENSATION p.25/40

2 Modelling Shape and Motion Unlike Kalman filters, CONDENSATION allows for arbitrary probability density functions The goal is to calculate, the probability of object state given the history of feature measurements 0 CONDENSATION p.26/40

Modelling Shape and Motion Probability density propagation with CONDENSATION. CONDENSATION p.27/40

2 2 2 4 2 2 " Modelling Shape and Motion Assumptions Markov assumption Assume the current object state distribution is dependent only on the previous state - CONDENSATION actually uses a second-order model, but the idea is the same Time-independence of the observation density "4 3 465-2 0 87 " This allows the observation density to be a static function CONDENSATION p.28/40

2 " <. 2? " = : " 9?? " Modelling Shape and Motion Measurement Want to determine the observation density a hypothesized object state After many assumptions, given / > B A@ $ > > ; where is the curve represented by shape parameter, is the closest feature to variance constant > >, > The observation process is ugly, but can be done is a CONDENSATION p.29/40

2 " " Factored Sampling General Idea Factored sampling is a method for approximating probability densities The problem is to find, which represents all knowledge about deducible from the data CONDENSATION p.30/40

2 " C 2 " 2 " H G D E!!! E J4 Factored Sampling General Idea Bayes rule can be used in principle: In practice, however, this cannot be solved in closed form Factored sampling generates a random variate that approximates Two parts A sample set Weights,.FE / H GI, corresponding to each sample H G4 CONDENSATION p.3/40

H G D E!!! E 2 E " J4 2 " Factored Sampling Details First, the set from the prior density.fe / H GI Each element is assigned a weight to the observation density H G4 is sampled randomly H G4 J4 proportional H G4 E K H GL E MK I L 5 K CONDENSATION p.32/40

Factored Sampling One dimensional case CONDENSATION p.33/40

2 2 The Algorithm The goal is to estimate the state probability density given and " " Since the state density is approximated as a sample set, the algorithm needs to generate a new sample set at each time CONDENSATION follows the same basic iterative steps as a Kalman filter: prediction, measurement, and assimilation N CONDENSATION p.34/40

O!!! P J O H E 2 E 2 J The Algorithm, samples from time Have H G4 H G4 / O.FE D D Want to construct, SR Q For according to from GUT ( D. Select a sample probability J H GUT ( H GUT E by sampling from 2. Predict 3. Measure and weight the new position according to image features, " H G T H G T E " CONDENSATION p.35/40

The Algorithm One time-step CONDENSATION p.36/40

Benefits Support for multi-modal probability distributions Multi-object support, less likely to lose track of objects Robustness Computationally efficient Adjustable cost/performance ratio (by changing the number of samples used) Relatively simple CONDENSATION p.37/40

Limitations Initialization Object model needs to be known in advance Iterative - still prone to getting stuck CONDENSATION p.38/40

Demonstration Hand Leaf Pointing finger Dancing girl Drawing CONDENSATION p.39/40

Conclusion Got questions? That s all folks CONDENSATION p.40/40