Lecture 22: Graph-SLAM (2)
|
|
- Kelly Neal
- 5 years ago
- Views:
Transcription
1 Lecture 22: Graph-SLAM (2) Dr. J.B. Hayet CENTRO DE INVESTIGACIÓN EN MATEMÁTICAS Abril 2014 J.B. Hayet Probabilistic robotics Abril / 35
2 Outline 1 Data association in Graph-SLAM 2 Improvements and examples J.B. Hayet Probabilistic robotics Abril / 35
3 Full SLAM In that case the state vector is X 0:t that combines all positions in the trajectory and the positions of the map elements X 0:t = R 0 R 1... R t M y X t = ( Rt M ). The posterior we want to estimate is p(x 0:t Z 1:t U 1:t C 1:t ). J.B. Hayet Probabilistic robotics Abril / 35
4 Graph-SLAM [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
5 Graph-SLAM [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
6 Graph-SLAM [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
7 Graph-SLAM: general algorithm 1 Inicializar µ. 2 Repeat until convergence: (ξ Ω) Linearize() ( ξ Ω) Reduce() (µ Σ 0:t ) Resolve() 3 Returns µ y Σ R0:t R 0:t. J.B. Hayet Probabilistic robotics Abril / 35
8 Outline Data association in Graph-SLAM 1 Data association in Graph-SLAM 2 Improvements and examples J.B. Hayet Probabilistic robotics Abril / 35
9 Data association in Graph-SLAM Data association In EKF-SLAM for data association: check for all i (observations) which map element j among all of the map best corresponds (or if you need to create a new one); In Graph-SLAM the idea is to start with unique correspondences (i.e. initially as many landmarks as observations) and to merge the ones that correspond to identical physical elements two by two. Hence at the core of this strategy is to model the distribution of m j m k for any (j k) to test the hypothesis that the 2 landmarks are the same. J.B. Hayet Probabilistic robotics Abril / 35
10 Data association in Graph-SLAM Data association It is a batch problem hence one can take advantage of some specificities: one can consider the observations in any order and even any sub-set of the observations at each step; the computation of correspondences and the one of the map are mixed; to choose among two associations and merge them will lead to different maps different posterior correspondences... a priori associations could be un-done. J.B. Hayet Probabilistic robotics Abril / 35
11 Data association in Graph-SLAM Data association For any pair of characteristics j and k in the map we can estimate the joint distribution of m j and m k : p(m j m k Z 1:t U 1:t C 1:t ) = R t... R 1 p(m j m k R 1:t Z 1:t U 1:t C 1:t )p(r 1:t Z 1:t U 1:t C 1:t )dr 1...dR t. As we saw it this distribution is Gaussian as a marginal of the joint distribution (ξ Ω). J.B. Hayet Probabilistic robotics Abril / 35
12 Data association in Graph-SLAM Data association Let τ(j) and τ(k) be the set of poses at which landmark j or k are seen τ(j k) = τ(j) τ(k) Now let us remind that the output of the algorithm is the mean µ over trajectories and map elements. J.B. Hayet Probabilistic robotics Abril / 35
13 Data association in Graph-SLAM Data association With a Gaussian of joint state vector (x y) represented in the form of information matrix/vector ( ) Ωxx Ω Ω = xy y ξ = Ω yx Ω yy the marginal over x can be written through: and ( ξx Ω xx = Ω xx Ω xy Ω 1 yy Ω yx ξ x = ξ x Ω xy Ω 1 yy ξ y. ξ y ) J.B. Hayet Probabilistic robotics Abril / 35
14 Data association in Graph-SLAM Data association Let ( ξ mj m k Ω mj m k ) be the parameters of the marginal over the landmarks positions j and k we can deduce from the marginalization formula: Ω mj m k = Ω mj m k m j m k Ω mj m k R τ (jk)σ Rτ (jk)r τ (jk)ω Rτ (jk)m j m k where Σ Rτ (jk)r τ (jk) is the sub-matrix of the covariance matrix over trajectories Σ R0:t R 0:t (output of the algorithm). J.B. Hayet Probabilistic robotics Abril / 35
15 Data association in Graph-SLAM Data association Ω Rτ(jk) R τ(jk) Ω mj m k R τ(jk) R τ(jk)τ(jk) Ω mj m k j k J.B. Hayet Probabilistic robotics Abril / 35
16 Data association in Graph-SLAM Data association Finally the joint distribution over the positions of the landmarks j and k has the following parameters: and Ω mj m k = Ω mj m k m j m k Ω mj m k R τ (jk)σ Rτ (jk)r τ (jk)ω Rτ (jk)m j m k ξ mj m k = Ωmj m k m j m k µ mj mk = Ω mj m k m j m k Ω 1 m j m k m j m k (ξ mj m k + Ω mj m k R τ(jk) µ τ(jk) ) J.B. Hayet Probabilistic robotics Abril / 35
17 Data association in Graph-SLAM Data association We deduce the expression of the joint distribution : p(m j m k Z 1:t U 1:t C 1:t ) e 1 2 m j m k T Ω mj m k m j m k m + j m k T ξ mj m k. Now we can examine the distribution of jk = m j m k that we can link easily to the joint one: ( ) ( ) ( ) ( ) mj m k 1 1 mj mj = = J. m j + m k 1 1 m k m k J.B. Hayet Probabilistic robotics Abril / 35
18 Data association in Graph-SLAM Data association ( ) mj m From the formula k has also a Gaussian distribution m j + m k ( ) ( ) µmj µ mean: mk µmj = J µ mj + µ mk µ mk covariance JΣ mj m k J T In particular the marginal of m j m k has covariance: ( 1 1 ) T Σ mj mk ( 1 1 ). J.B. Hayet Probabilistic robotics Abril / 35
19 Data association in Graph-SLAM Data association Its inverse (check it!): Ω jk = 1 4 ( 1 1 ) T Σ 1 m j m k ( and in terms of what we have already computed: Ω jk = 1 4 ( 1 1 ) T Ωmj mk 1 1 ( 1 1 ). ). J.B. Hayet Probabilistic robotics Abril / 35
20 Data association in Graph-SLAM Data association In the classic form of moments : p( jk Z 1:t U 1:t C 1:t ) = 1 det(2πω 1 jk ) e 1 2 ( µ ) T Ω jk ( µ ) and the test to be done is: which can be written: p( jk = 0 Z 1:t U 1:t C 1:t ) 1 e 1 2 µt Ω jk µ. det(2πω 1 jk ) J.B. Hayet Probabilistic robotics Abril / 35
21 Data association in Graph-SLAM Data association: modified algorithm Correspondence test j k given Ω ξ (joint distribution) µ Σ 0:t : Ω mj m k = Ω mj m k m j m k Ω mj m k R τ (jk)σ Rτ (jk)r τ (jk)ω Rτ (jk)m j m k ( ) T ( ) 1 1 Ω jk = 1 4 Ω 1 mj mk 1 µ = µ j µ k 1 π jk = e 1 2 µt Ω jk µ det(2πω 1 ) jk Si π jk > τ returns true otherwise returns false. J.B. Hayet Probabilistic robotics Abril / 35
22 Data association in Graph-SLAM Data association: modified algorithm Initialize correspondences c 1:t in a unique way 1 Inicializar µ. 2 (ξ Ω) Linearize() 3 ( ξ Ω) Reduce() 4 (µ Σ 0:t ) Resolve() 5 repeat until there is no more mergeable pairs j k: if TestCorrespondence(j k ξ Ω µ Σ 0:t ) re-estimate trajectories and map with the modified correspondences: i/c i t = k c i t = j. J.B. Hayet Probabilistic robotics Abril / 35
23 Data association in Graph-SLAM Data association: modified algorithm Once the correspondences vector is corrected restart everything with the modified c 1:t : 1 (ξ Ω) Linearize() 2 ( ξ Ω) Reduce() 3 (µ Σ 0:t ) Resolve() J.B. Hayet Probabilistic robotics Abril / 35
24 Data association in Graph-SLAM Data association: limits Greedy algorithm because again the problem is combinatorial (explore all possible associations) By starting by unique correspondences we then have to check merges two by two this is not efficient. It does not take into account the fact that at some t one landmark can only be seen through one observation not two i j c i t c j t with the algorithm above we could have cases of associations for the same time t. But it is simple to handle this constraint. J.B. Hayet Probabilistic robotics Abril / 35
25 Data association in Graph-SLAM Data association: limits In the exploration of data associations the order counts: by starting with one fusion we may lose the track of the good associations set. Negative information is not considered (absence of a landmark that should have been seen); We do not use the possibility to remove a correspondence... There are techniques dedicated to these problems. J.B. Hayet Probabilistic robotics Abril / 35
26 Data association in Graph-SLAM Data association: limits This algorithm is essentially a theoretical base; many extensions could be proposed: merge features since the beginning when it is very likely that they correspond to the same element; use sub-maps built with other techniques locally efficient (EKF occupancy grids... ) then use Graph-SLAM between sub-maps. J.B. Hayet Probabilistic robotics Abril / 35
27 Data association in Graph-SLAM Graph-SLAM and other methods Graph-SLAM has a lot in common with the technique of Estrada et al. (cf. lecture 20) to stick sub-maps together except that here one handles information matrices In the same way in the most advanced scan-matching techniques use generally graph-based representations to manage the loop closure problem... J.B. Hayet Probabilistic robotics Abril / 35
28 Outline Improvements and examples 1 Data association in Graph-SLAM 2 Improvements and examples J.B. Hayet Probabilistic robotics Abril / 35
29 Improvements and examples Example: mine exploration Example from the book: mapping of abandoned mines... Network of corridors hundreds of meters long... Algorithm: 1 construction of an occupancy grid with scan matching (see previous lecture) 2 division of the map in sub-maps of 5m 3 construction of a graph where each node corresponds to a sub-map (pose node in Graph-SLAM) 4 association test for identical data (map correlation) and re-estimation of the map. J.B. Hayet Probabilistic robotics Abril / 35
30 Improvements and examples Example: mine exploration [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
31 Improvements and examples Example: mine exploration [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
32 Improvements and examples Example: mine exploration [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
33 Improvements and examples Example: mine exploration [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
34 Improvements and examples Example: mine exploration [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
35 Improvements and examples Example: mine exploration [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
36 Improvements and examples Optimization-based variants An improvement point resides in the optimization; what we could find is the map/trajectory maximizing the posterior i.e. minimizing: l 0:t = cste. + R0 T Ω 0 R 0 + t 1 (R 2 τ g(r τ 1 U τ )) T Σ 1 R (R τ g(r τ 1 U τ ))+ τ=1 t o τ (Zτ i h(x τ Cτ)) i T Σ 1 M (Z τ i h(x τ Cτ)) i 1 2 τ=1 i=1. The basic algorithm linearizes g and h... but it has the advantage of estimating the covariance on trajectories. J.B. Hayet Probabilistic robotics Abril / 35
37 Improvements and examples Optimization-based variants Instead of sequences linearizations/reductions/optimizations this problem can be handled by classical optimization methods to determine the MAP (maximum a posteriori) : gradient descent; Levenberg-Marquardt; conjugated gradient... Most of the literature is based on this type of approach. J.B. Hayet Probabilistic robotics Abril / 35
38 Improvements and examples Optimization-based variants They allow to handle much larger dimensions (examples up to 10 8 features... ); They are faster in general than the original algorithm; They are sensible to local minima... They just output the mode (MAP). J.B. Hayet Probabilistic robotics Abril / 35
39 Improvements and examples Optimization-based variants [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
40 Improvements and examples Optimization-based variants [From Probabilistic Robotics MIT Press] J.B. Hayet Probabilistic robotics Abril / 35
Simultaneous Localization and Mapping
Simultaneous Localization and Mapping Miroslav Kulich Intelligent and Mobile Robotics Group Czech Institute of Informatics, Robotics and Cybernetics Czech Technical University in Prague Winter semester
More informationL11. EKF SLAM: PART I. NA568 Mobile Robotics: Methods & Algorithms
L11. EKF SLAM: PART I NA568 Mobile Robotics: Methods & Algorithms Today s Topic EKF Feature-Based SLAM State Representation Process / Observation Models Landmark Initialization Robot-Landmark Correlation
More informationProbabilistic Fundamentals in Robotics. DAUIN Politecnico di Torino July 2010
Probabilistic Fundamentals in Robotics Probabilistic Models of Mobile Robots Robotic mapping Basilio Bona DAUIN Politecnico di Torino July 2010 Course Outline Basic mathematical framework Probabilistic
More informationIntroduction to Mobile Robotics SLAM: Simultaneous Localization and Mapping
Introduction to Mobile Robotics SLAM: Simultaneous Localization and Mapping Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz What is SLAM? Estimate the pose of a robot and the map of the environment
More informationSLAM for Ship Hull Inspection using Exactly Sparse Extended Information Filters
SLAM for Ship Hull Inspection using Exactly Sparse Extended Information Filters Matthew Walter 1,2, Franz Hover 1, & John Leonard 1,2 Massachusetts Institute of Technology 1 Department of Mechanical Engineering
More informationParticle Filters; Simultaneous Localization and Mapping (Intelligent Autonomous Robotics) Subramanian Ramamoorthy School of Informatics
Particle Filters; Simultaneous Localization and Mapping (Intelligent Autonomous Robotics) Subramanian Ramamoorthy School of Informatics Recap: State Estimation using Kalman Filter Project state and error
More informationMobile Robot Localization
Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations
More informationSLAM Techniques and Algorithms. Jack Collier. Canada. Recherche et développement pour la défense Canada. Defence Research and Development Canada
SLAM Techniques and Algorithms Jack Collier Defence Research and Development Canada Recherche et développement pour la défense Canada Canada Goals What will we learn Gain an appreciation for what SLAM
More informationCS8803: Statistical Techniques in Robotics Byron Boots. Hilbert Space Embeddings
CS8803: Statistical Techniques in Robotics Byron Boots Hilbert Space Embeddings 1 Motivation CS8803: STR Hilbert Space Embeddings 2 Overview Multinomial Distributions Marginal, Joint, Conditional Sum,
More informationRobot Localization and Kalman Filters
Robot Localization and Kalman Filters Rudy Negenborn rudy@negenborn.net August 26, 2003 Outline Robot Localization Probabilistic Localization Kalman Filters Kalman Localization Kalman Localization with
More informationStat 521A Lecture 18 1
Stat 521A Lecture 18 1 Outline Cts and discrete variables (14.1) Gaussian networks (14.2) Conditional Gaussian networks (14.3) Non-linear Gaussian networks (14.4) Sampling (14.5) 2 Hybrid networks A hybrid
More informationEKF and SLAM. McGill COMP 765 Sept 18 th, 2017
EKF and SLAM McGill COMP 765 Sept 18 th, 2017 Outline News and information Instructions for paper presentations Continue on Kalman filter: EKF and extension to mapping Example of a real mapping system:
More informationRobotics 2 Target Tracking. Kai Arras, Cyrill Stachniss, Maren Bennewitz, Wolfram Burgard
Robotics 2 Target Tracking Kai Arras, Cyrill Stachniss, Maren Bennewitz, Wolfram Burgard Slides by Kai Arras, Gian Diego Tipaldi, v.1.1, Jan 2012 Chapter Contents Target Tracking Overview Applications
More informationMobile Robot Localization
Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations
More informationMultivariate Gaussians. Sargur Srihari
Multivariate Gaussians Sargur srihari@cedar.buffalo.edu 1 Topics 1. Multivariate Gaussian: Basic Parameterization 2. Covariance and Information Form 3. Operations on Gaussians 4. Independencies in Gaussians
More informationCIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions
CIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions December 14, 2016 Questions Throughout the following questions we will assume that x t is the state vector at time t, z t is the
More informationRobotics. Mobile Robotics. Marc Toussaint U Stuttgart
Robotics Mobile Robotics State estimation, Bayes filter, odometry, particle filter, Kalman filter, SLAM, joint Bayes filter, EKF SLAM, particle SLAM, graph-based SLAM Marc Toussaint U Stuttgart DARPA Grand
More informationTSRT14: Sensor Fusion Lecture 9
TSRT14: Sensor Fusion Lecture 9 Simultaneous localization and mapping (SLAM) Gustaf Hendeby gustaf.hendeby@liu.se TSRT14 Lecture 9 Gustaf Hendeby Spring 2018 1 / 28 Le 9: simultaneous localization and
More informationIntroduction to Mobile Robotics Information Gain-Based Exploration. Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Giorgio Grisetti, Kai Arras
Introduction to Mobile Robotics Information Gain-Based Exploration Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Giorgio Grisetti, Kai Arras 1 Tasks of Mobile Robots mapping SLAM localization integrated
More informationRandom Process. Random Process. Random Process. Introduction to Random Processes
Random Process A random variable is a function X(e) that maps the set of experiment outcomes to the set of numbers. A random process is a rule that maps every outcome e of an experiment to a function X(t,
More informationAUTOMOTIVE ENVIRONMENT SENSORS
AUTOMOTIVE ENVIRONMENT SENSORS Lecture 5. Localization BME KÖZLEKEDÉSMÉRNÖKI ÉS JÁRMŰMÉRNÖKI KAR 32708-2/2017/INTFIN SZÁMÚ EMMI ÁLTAL TÁMOGATOTT TANANYAG Related concepts Concepts related to vehicles moving
More informationRao-Blackwellized Particle Filter for Multiple Target Tracking
Rao-Blackwellized Particle Filter for Multiple Target Tracking Simo Särkkä, Aki Vehtari, Jouko Lampinen Helsinki University of Technology, Finland Abstract In this article we propose a new Rao-Blackwellized
More informationLecture 7: Optimal Smoothing
Department of Biomedical Engineering and Computational Science Aalto University March 17, 2011 Contents 1 What is Optimal Smoothing? 2 Bayesian Optimal Smoothing Equations 3 Rauch-Tung-Striebel Smoother
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationIntroduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf
1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Lior Wolf 2014-15 We know that X ~ B(n,p), but we do not know p. We get a random sample from X, a
More informationBayes Filter Reminder. Kalman Filter Localization. Properties of Gaussians. Gaussians. Prediction. Correction. σ 2. Univariate. 1 2πσ e.
Kalman Filter Localization Bayes Filter Reminder Prediction Correction Gaussians p(x) ~ N(µ,σ 2 ) : Properties of Gaussians Univariate p(x) = 1 1 2πσ e 2 (x µ) 2 σ 2 µ Univariate -σ σ Multivariate µ Multivariate
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationL06. LINEAR KALMAN FILTERS. NA568 Mobile Robotics: Methods & Algorithms
L06. LINEAR KALMAN FILTERS NA568 Mobile Robotics: Methods & Algorithms 2 PS2 is out! Landmark-based Localization: EKF, UKF, PF Today s Lecture Minimum Mean Square Error (MMSE) Linear Kalman Filter Gaussian
More informationUniversität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Bayesian Learning. Tobias Scheffer, Niels Landwehr
Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning Tobias Scheffer, Niels Landwehr Remember: Normal Distribution Distribution over x. Density function with parameters
More informationProbabilistic Fundamentals in Robotics
Probabilistic Fundamentals in Robotics Probabilistic Models of Mobile Robots Robot localization Basilio Bona DAUIN Politecnico di Torino June 2011 Course Outline Basic mathematical framework Probabilistic
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationOutline. Supervised Learning. Hong Chang. Institute of Computing Technology, Chinese Academy of Sciences. Machine Learning Methods (Fall 2012)
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Linear Models for Regression Linear Regression Probabilistic Interpretation
More informationParametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012
Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood
More informationDistributed Robust Localization from Arbitrary Initial Poses via EM and Model Selection
1 Distributed Robust Localization from Arbitrary Initial Poses via EM and Model Selection Vadim Indelman Collaborators: Erik Nelson, Jing Dong, Nathan Michael and Frank Dellaert IACAS, February 15 Collaborative
More informationProbabilistic Fundamentals in Robotics. DAUIN Politecnico di Torino July 2010
Probabilistic Fundamentals in Robotics Probabilistic Models of Mobile Robots Robot localization Basilio Bona DAUIN Politecnico di Torino July 2010 Course Outline Basic mathematical framework Probabilistic
More informationSimultaneous Localization and Mapping (SLAM) Corso di Robotica Prof. Davide Brugali Università degli Studi di Bergamo
Simultaneous Localization and Mapping (SLAM) Corso di Robotica Prof. Davide Brugali Università degli Studi di Bergamo Introduction SLAM asks the following question: Is it possible for an autonomous vehicle
More informationArtificial Intelligence
ICS461 Fall 2010 Nancy E. Reed nreed@hawaii.edu 1 Lecture #14B Outline Inference in Bayesian Networks Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic
More informationIntro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation
Lecture 15. Pattern Classification (I): Statistical Formulation Outline Statistical Pattern Recognition Maximum Posterior Probability (MAP) Classifier Maximum Likelihood (ML) Classifier K-Nearest Neighbor
More informationOverfitting, Bias / Variance Analysis
Overfitting, Bias / Variance Analysis Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 8, 207 / 40 Outline Administration 2 Review of last lecture 3 Basic
More informationLecture 13 and 14: Bayesian estimation theory
1 Lecture 13 and 14: Bayesian estimation theory Spring 2012 - EE 194 Networked estimation and control (Prof. Khan) March 26 2012 I. BAYESIAN ESTIMATORS Mother Nature conducts a random experiment that generates
More informationKalman Filter. Predict: Update: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q
Kalman Filter Kalman Filter Predict: x k k 1 = F k x k 1 k 1 + B k u k P k k 1 = F k P k 1 k 1 F T k + Q Update: K = P k k 1 Hk T (H k P k k 1 Hk T + R) 1 x k k = x k k 1 + K(z k H k x k k 1 ) P k k =(I
More informationA Probability Review
A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in
More information9 Multi-Model State Estimation
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 9 Multi-Model State
More informationBayesian Networks Inference with Probabilistic Graphical Models
4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning
More informationCHAPTER-17. Decision Tree Induction
CHAPTER-17 Decision Tree Induction 17.1 Introduction 17.2 Attribute selection measure 17.3 Tree Pruning 17.4 Extracting Classification Rules from Decision Trees 17.5 Bayesian Classification 17.6 Bayes
More informationProbabilistic Fundamentals in Robotics. DAUIN Politecnico di Torino July 2010
Probabilistic Fundamentals in Robotics Gaussian Filters Basilio Bona DAUIN Politecnico di Torino July 2010 Course Outline Basic mathematical framework Probabilistic models of mobile robots Mobile robot
More informationIntroduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf
1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Yishay Mansour, Lior Wolf 2013-14 We know that X ~ B(n,p), but we do not know p. We get a random sample
More informationManipulators. Robotics. Outline. Non-holonomic robots. Sensors. Mobile Robots
Manipulators P obotics Configuration of robot specified by 6 numbers 6 degrees of freedom (DOF) 6 is the minimum number required to position end-effector arbitrarily. For dynamical systems, add velocity
More informationGaussian processes and bayesian optimization Stanisław Jastrzębski. kudkudak.github.io kudkudak
Gaussian processes and bayesian optimization Stanisław Jastrzębski kudkudak.github.io kudkudak Plan Goal: talk about modern hyperparameter optimization algorithms Bayes reminder: equivalent linear regression
More informationProblems on Discrete & Continuous R.Vs
013 SUBJECT NAME SUBJECT CODE MATERIAL NAME MATERIAL CODE : Probability & Random Process : MA 61 : University Questions : SKMA1004 Name of the Student: Branch: Unit I (Random Variables) Problems on Discrete
More informationTDA231. Logistic regression
TDA231 Devdatt Dubhashi dubhashi@chalmers.se Dept. of Computer Science and Engg. Chalmers University February 19, 2016 Some data 5 x2 0 5 5 0 5 x 1 In the Bayes classifier, we built a model of each class
More informationLeast Squares Regression
CIS 50: Machine Learning Spring 08: Lecture 4 Least Squares Regression Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may not cover all the
More informationECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering
ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:
More informationLast Time. Today. Bayesian Learning. The Distributions We Love. CSE 446 Gaussian Naïve Bayes & Logistic Regression
CSE 446 Gaussian Naïve Bayes & Logistic Regression Winter 22 Dan Weld Learning Gaussians Naïve Bayes Last Time Gaussians Naïve Bayes Logistic Regression Today Some slides from Carlos Guestrin, Luke Zettlemoyer
More informationMark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.
University of Cambridge Engineering Part IIB & EIST Part II Paper I0: Advanced Pattern Processing Handouts 4 & 5: Multi-Layer Perceptron: Introduction and Training x y (x) Inputs x 2 y (x) 2 Outputs x
More informationROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino
ROBOTICS 01PEEQW Basilio Bona DAUIN Politecnico di Torino Probabilistic Fundamentals in Robotics Gaussian Filters Course Outline Basic mathematical framework Probabilistic models of mobile robots Mobile
More informationTSRT14: Sensor Fusion Lecture 8
TSRT14: Sensor Fusion Lecture 8 Particle filter theory Marginalized particle filter Gustaf Hendeby gustaf.hendeby@liu.se TSRT14 Lecture 8 Gustaf Hendeby Spring 2018 1 / 25 Le 8: particle filter theory,
More informationE190Q Lecture 11 Autonomous Robot Navigation
E190Q Lecture 11 Autonomous Robot Navigation Instructor: Chris Clark Semester: Spring 013 1 Figures courtesy of Siegwart & Nourbakhsh Control Structures Planning Based Control Prior Knowledge Operator
More informationRobotics 2 Data Association. Giorgio Grisetti, Cyrill Stachniss, Kai Arras, Wolfram Burgard
Robotics 2 Data Association Giorgio Grisetti, Cyrill Stachniss, Kai Arras, Wolfram Burgard Data Association Data association is the process of associating uncertain measurements to known tracks. Problem
More informationGaussian with mean ( µ ) and standard deviation ( σ)
Slide from Pieter Abbeel Gaussian with mean ( µ ) and standard deviation ( σ) 10/6/16 CSE-571: Robotics X ~ N( µ, σ ) Y ~ N( aµ + b, a σ ) Y = ax + b + + + + 1 1 1 1 1 1 1 1 1 1, ~ ) ( ) ( ), ( ~ ), (
More informationLinear Models for Classification
Linear Models for Classification Henrik I Christensen Robotics & Intelligent Machines @ GT Georgia Institute of Technology, Atlanta, GA 30332-0280 hic@cc.gatech.edu Henrik I Christensen (RIM@GT) Linear
More informationCovariance to PCA. CS 510 Lecture #14 February 23, 2018
Covariance to PCA CS 510 Lecture 14 February 23, 2018 Overview: Goal Assume you have a gallery (database) of images, and a probe (test) image. The goal is to find the database image that is most similar
More informationGradient Descent and Implementation Solving the Euler-Lagrange Equations in Practice
1 Lecture Notes, HCI, 4.1.211 Chapter 2 Gradient Descent and Implementation Solving the Euler-Lagrange Equations in Practice Bastian Goldlücke Computer Vision Group Technical University of Munich 2 Bastian
More informationAbsolute map-based localization for a planetary rover
Absolute map-based localization for a planetary rover Bach Van Pham, Artur Maligo and Simon Lacroix LAAS/CNRS, Toulouse Work developed within the ESA founded Startiger activity Seeker Outline" On the importance
More informationMassachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Problem Set 3 Issued: Thursday, September 25, 2014 Due: Thursday,
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationCMU-Q Lecture 24:
CMU-Q 15-381 Lecture 24: Supervised Learning 2 Teacher: Gianni A. Di Caro SUPERVISED LEARNING Hypotheses space Hypothesis function Labeled Given Errors Performance criteria Given a collection of input
More informationK-Means and Gaussian Mixture Models
K-Means and Gaussian Mixture Models David Rosenberg New York University October 29, 2016 David Rosenberg (New York University) DS-GA 1003 October 29, 2016 1 / 42 K-Means Clustering K-Means Clustering David
More informationIntroduction to Probabilistic Reasoning. Image credit: NASA. Assignment
Introduction to Probabilistic Reasoning Brian C. Williams 16.410/16.413 November 17 th, 2010 11/17/10 copyright Brian Williams, 2005-10 1 Brian C. Williams, copyright 2000-09 Image credit: NASA. Assignment
More informationQuantitative Information Flow. Lecture 7
Quantitative Information Flow Lecture 7 1 The basic model: Systems = Information-Theoretic channels Secret Information Observables s1 o1... System... sm on Input Output 2 Probabilistic systems are noisy
More informationChapter 4 Multiple Random Variables
Review for the previous lecture Definition: n-dimensional random vector, joint pmf (pdf), marginal pmf (pdf) Theorem: How to calculate marginal pmf (pdf) given joint pmf (pdf) Example: How to calculate
More informationProbabilistic Graphical Models
School of Computer Science Probabilistic Graphical Models Variational Inference II: Mean Field Method and Variational Principle Junming Yin Lecture 15, March 7, 2012 X 1 X 1 X 1 X 1 X 2 X 3 X 2 X 2 X 3
More informationFundamentals of Unconstrained Optimization
dalmau@cimat.mx Centro de Investigación en Matemáticas CIMAT A.C. Mexico Enero 2016 Outline Introduction 1 Introduction 2 3 4 Optimization Problem min f (x) x Ω where f (x) is a real-valued function The
More informationLecture : Probabilistic Machine Learning
Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning
More informationCovariance. if X, Y are independent
Review: probability Monty Hall, weighted dice Frequentist v. Bayesian Independence Expectations, conditional expectations Exp. & independence; linearity of exp. Estimator (RV computed from sample) law
More information01 Probability Theory and Statistics Review
NAVARCH/EECS 568, ROB 530 - Winter 2018 01 Probability Theory and Statistics Review Maani Ghaffari January 08, 2018 Last Time: Bayes Filters Given: Stream of observations z 1:t and action data u 1:t Sensor/measurement
More informationKalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein
Kalman filtering and friends: Inference in time series models Herke van Hoof slides mostly by Michael Rubinstein Problem overview Goal Estimate most probable state at time k using measurement up to time
More informationExact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = =
Exact Inference I Mark Peot In this lecture we will look at issues associated with exact inference 10 Queries The objective of probabilistic inference is to compute a joint distribution of a set of query
More informationLogistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu
Logistic Regression Review 10-601 Fall 2012 Recitation September 25, 2012 TA: Selen Uguroglu!1 Outline Decision Theory Logistic regression Goal Loss function Inference Gradient Descent!2 Training Data
More informationGibbs Fields & Markov Random Fields
Statistical Techniques in Robotics (16-831, F10) Lecture#7 (Tuesday September 21) Gibbs Fields & Markov Random Fields Lecturer: Drew Bagnell Scribe: Bradford Neuman 1 1 Gibbs Fields Like a Bayes Net, a
More informationIntroduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization
Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Kai Arras 1 Motivation Recall: Discrete filter Discretize the
More informationMidterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas
Midterm Review CS 6375: Machine Learning Vibhav Gogate The University of Texas at Dallas Machine Learning Supervised Learning Unsupervised Learning Reinforcement Learning Parametric Y Continuous Non-parametric
More informationProbabilistic Graphical Networks: Definitions and Basic Results
This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical
More informationDynamic models 1 Kalman filters, linearization,
Koller & Friedman: Chapter 16 Jordan: Chapters 13, 15 Uri Lerner s Thesis: Chapters 3,9 Dynamic models 1 Kalman filters, linearization, Switching KFs, Assumed density filters Probabilistic Graphical Models
More informationUsing the Kalman Filter for SLAM AIMS 2015
Using the Kalman Filter for SLAM AIMS 2015 Contents Trivial Kinematics Rapid sweep over localisation and mapping (components of SLAM) Basic EKF Feature Based SLAM Feature types and representations Implementation
More informationMachine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io
Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem
More informationLecture 9: PGM Learning
13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and
More informationRobotics 2 Target Tracking. Giorgio Grisetti, Cyrill Stachniss, Kai Arras, Wolfram Burgard
Robotics 2 Target Tracking Giorgio Grisetti, Cyrill Stachniss, Kai Arras, Wolfram Burgard Linear Dynamical System (LDS) Stochastic process governed by is the state vector is the input vector is the process
More informationCS 532: 3D Computer Vision 6 th Set of Notes
1 CS 532: 3D Computer Vision 6 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Lecture Outline Intro to Covariance
More informationProbabilistic Machine Learning. Industrial AI Lab.
Probabilistic Machine Learning Industrial AI Lab. Probabilistic Linear Regression Outline Probabilistic Classification Probabilistic Clustering Probabilistic Dimension Reduction 2 Probabilistic Linear
More informationClustering using Mixture Models
Clustering using Mixture Models The full posterior of the Gaussian Mixture Model is p(x, Z, µ,, ) =p(x Z, µ, )p(z )p( )p(µ, ) data likelihood (Gaussian) correspondence prob. (Multinomial) mixture prior
More informationLecture 11: Probability Distributions and Parameter Estimation
Intelligent Data Analysis and Probabilistic Inference Lecture 11: Probability Distributions and Parameter Estimation Recommended reading: Bishop: Chapters 1.2, 2.1 2.3.4, Appendix B Duncan Gillies and
More informationLecture 2: Simple Classifiers
CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 2: Simple Classifiers Slides based on Rich Zemel s All lecture slides will be available on the course website: www.cs.toronto.edu/~jessebett/csc412
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should
More informationLinear Dynamical Systems (Kalman filter)
Linear Dynamical Systems (Kalman filter) (a) Overview of HMMs (b) From HMMs to Linear Dynamical Systems (LDS) 1 Markov Chains with Discrete Random Variables x 1 x 2 x 3 x T Let s assume we have discrete
More informationAutonomous Mobile Robot Design
Autonomous Mobile Robot Design Topic: Extended Kalman Filter Dr. Kostas Alexis (CSE) These slides relied on the lectures from C. Stachniss, J. Sturm and the book Probabilistic Robotics from Thurn et al.
More informationStochastic Generative Hashing
Stochastic Generative Hashing B. Dai 1, R. Guo 2, S. Kumar 2, N. He 3 and L. Song 1 1 Georgia Institute of Technology, 2 Google Research, NYC, 3 University of Illinois at Urbana-Champaign Discussion by
More informationPerhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.
Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage
More informationLeast Squares Regression
E0 70 Machine Learning Lecture 4 Jan 7, 03) Least Squares Regression Lecturer: Shivani Agarwal Disclaimer: These notes are a brief summary of the topics covered in the lecture. They are not a substitute
More informationCS242: Probabilistic Graphical Models Lecture 4A: MAP Estimation & Graph Structure Learning
CS242: Probabilistic Graphical Models Lecture 4A: MAP Estimation & Graph Structure Learning Professor Erik Sudderth Brown University Computer Science October 4, 2016 Some figures and materials courtesy
More information