Spectral learning algorithms for dynamical systems

Size: px
Start display at page:

Download "Spectral learning algorithms for dynamical systems"

Transcription

1 Spectral learning algorithms for dynamical systems Geoff Gordon Machine Learning Department Carnegie Mellon University joint work with Byron Boots, Sajid Siddiqi, Le Song, Alex Smola

2 2 What s out there? ot-2 ot-1 ot ot+1 ot+2 video frames, as vectors of pixel intensities

3 2 What s out there? ot-2 ot-1 ot ot+1 ot+2 video frames, as vectors of pixel intensities

4 A dynamical system Past Future ot-2 ot-1 ot ot+1 ot+2 Given past observations from a partially observable system State Predict future observations 3

5 A dynamical system Past Future ot-2 ot-1 ot ot+1 ot+2 Given past observations from a partially observable system State Predict future observations 3

6 This talk General class of models for dynamical systems Fast, statistically consistent learning algorithm no local optima Includes many well-known models & algorithms as special cases Also includes new models & algorithms that give state-of-the-art performance on interesting tasks 4

7 Includes models hidden Markov model n-grams, regexes, k-order HMMs PSR, OOM, multiplicity automaton, RR-HMM LTI system Kalman filter, AR, ARMA Kernel versions and manifold versions of above 5

8 Includes algorithms Subspace identification for LTI systems recent extensions to HMMs, RR-HMMs Tomasi-Kanade structure from motion Principal components analysis (e.g., eigenfaces); Laplacian eigenmaps New algorithms for learning PSRs, OOMs, etc. A new way to reduce noise in manifold learning A new range-only SLAM algorithm 6

9 Interesting applications Structure from motion Simultaneous localization and mapping Range-only SLAM SLAM from inertial sensors very simple vision-based SLAM (so far) Video textures Opponent modeling, option pricing, audio event classification, 7

10 Bayes filter (HMM, Kalman, etc.) ot-2 ot-1 ot ot+1 ot st-2 st-1 st st+1 st+2 Belief Pt(st) = P(st o1:t) Goal: given ot, update Pt-1(st-1) to Pt(st) Extend: Pt-1(st-1, ot, st) = Pt-1(st-1) P(st st-1) P(ot st) Marginalize: Pt-1(ot, st) =! Pt-1(st-1, ot, st) dst-1 Condition: Pt(st) = Pt-1(ot, st) / Pt-1(ot) 8

11 Bayes filter (HMM, Kalman, etc.) ot-2 ot-1 ot ot+1 ot st-2 st-1 st st+1 st+2 Belief Pt(st) = P(st o1:t) Goal: given ot, update Pt-1(st-1) to Pt(st) Extend: Pt-1(st-1, ot, st) = Pt-1(st-1) P(st st-1) P(ot st) Marginalize: Pt-1(ot, st) =! Pt-1(st-1, ot, st) dst-1 Condition: Pt(st) = Pt-1(ot, st) / Pt-1(ot) 8

12 nb: A common form for Bayes filters Find covariances as (linear) fns of previous state Σo(st-1) = E(!(ot)!(ot) T st-1) Σso(st-1) = E(st!(ot) st-1) uncentered covars; st &!(ot) include constant Linear regression to get current state -1 st = Σso Σo!(ot) Exact if discrete (HMM, PSR), Gaussian (Kalman, AR), RKHS w/ characteristic kernel [Fukumizu et al.] Approximates many more 9

13 Why this form? If Bayes filter takes (approximately) the above form, can design a simple spectral algorithm to identify the system Intuitions: predictive state rank bottleneck observable representation 10

14 Predictive state If Bayes filter takes above form, we can use a vector of predictions of observables as our state E(!(ot+k) st) = linear fn of st for big enough k, E([!(ot+1)!(ot+k)] st) = invertible linear fn of st so, take E([!(ot+1)!(ot+k)] st) as our state 11

15 12 Predictive state: example

16 12 Predictive state: example

17 12 Predictive state: example

18 12 Predictive state: example

19 13 Predictive state: minimal example P(s t s t 1 ) = = T P(o t s t 1 ) = OT = = O For big enough k, E([!(ot+1)!(ot+k)] st) = Wst invertible linear fn (if system observable)

20 13 Predictive state: minimal example P(s t s t 1 ) = = T P(o t s t 1 ) = OT = = O For big enough k, E([!(ot+1)!(ot+k)] st) = Wst invertible linear fn (if system observable)

21 13 Predictive state: minimal example P(s t s t 1 ) = = T P(o t s t 1 ) = OT = = O W For big enough k, E([!(ot+1)!(ot+k)] st) = Wst invertible linear fn (if system observable)

22 14 Predictive state: summary E([!(ot+1)!(ot+k)] st) is our state rep n interpretable observable a natural target for learning

23 14 Predictive state: summary E([!(ot+1)!(ot+k)] st) is our state rep n interpretable observable a natural target for learning To find st, predict!(ot+1)!(ot+k) from o1:t

24 Rank bottleneck predict data about past (many samples) state data about future (many samples) compress bottleneck expand Can find best rank-k bottleneck via matrix factorization spectral method 15

25 Finding a compact predictive state W Y " X linear function predicts Yt from Xt 16

26 Finding a compact predictive state W Y " X linear function predicts Yt from Xt 16

27 Finding a compact predictive state W Y " X linear function predicts Yt from Xt 16

28 Finding a compact predictive state W Y " X linear function predicts Yt from Xt 16

29 Finding a compact predictive state W Y " X linear function predicts Yt from Xt 16

30 Finding a compact predictive state W Y " X linear function predicts Yt from Xt 16

31 Finding a compact predictive state W rank(w) # k Y " X linear function predicts Yt from Xt 16

32 Observable representation Now that we have a compact predictive state, need to estimate fns Σo(st) and Σso(st) Insight: parameters are now observable: the problem is just to estimate some covariances from data to get Σo, regress (!(o) $!(o)) state to get Σso, regress (!(o) $ future+) state Past Future ot-2 ot-1 ot ot+1 ot+2 State Future+ 17

33 Algorithm Find compact predictive state: regress future past use any features of past/future, or a kernel, or even a learned kernel (manifold case) constrain rank of prediction weight matrix Extract model: regress (o $ future+) [predictive state] (o $ o) [predictive state] future+: use same rank-constrained basis as for predictive state 18

34 19 Does it work? Transitions X T Observations Simple HMM: 5 states, 7 observations N=300 thru N=900k Y T

35 Does it work? Model True P(o1, o2, o3) mean abs err (10 0 =predict uniform) 10 0 run 1 run 2 c/sqrt(n) training data 20

36 Discussion Impossibility learning DFA in poly time = breaking crypto primitives [Kearns & Valiant 89] so, clearly, we can t always be statistically efficient but, see McDonald [11], HKZ [09], us [09]: convergence depends on mixing rate Nonlinearity Bayes filter update is highly nonlinear in state (matrix inverse), even though we use a linear regression to identify the model nonlinearity is essential for expressiveness 21

37 Range-only SLAM Robot measures distance from current position to fixed beacons (e.g., time of flight or signal strength) may also have odometry Goal: recover robot path, landmark locations 22

38 Typical solution: EKF Prior After 1 range measurement to known landmark Bayes filter vs. EKF Problem: linear/gaussian approximation very bad

39 Typical solution: EKF Prior After 1 range measurement to known landmark Bayes filter vs. EKF Problem: linear/gaussian approximation very bad

40 24 Spectral solution (simple version) landmark 1, time step 2 d 2 11 d d 2 1T d 2 21 d d 2 2T d 2 N1 d 2 N2... d 2 NT

41 24 Spectral solution (simple version) 1/2 landmark 1, time step 2 d 2 11 d d 2 1T d 2 21 d d 2 2T d 2 N1 d 2 N2... d 2 NT = landmark position x, y 1 x 1 y 1 (x y 2 1)/2 1 x 2 y 2 (x y 2 2)/ x N y N (x 2 N + y2 N )/2 (u v1)/2 2 (u v2)/ (u 2 T + v2 T )/2 u 1 u 2... u T v 1 v 2... v T robot position u, v

42 25 Experiments (full version) autonomous lawn mower 60 dead reackoning Path true Path est. Path true Landmark est. Landmark B. Plaza 1 C. Plaza 2 60 Plaza 2 [Djugash & Singh, 2008] Plaza

43 Experiments (full version) urrent MAP dead reackoning Path true Path est. Path Method Plaza 1 Plaza 2 true Landmark Dead Reckoning (full path) est. Landmark 15.92m 27.28m [Djugash & Singh, 2008] Cartesian EKF (last 10%) 0.94m 0.92m Plaza B. 3OD]D C. Plaza 2 Fig. 4. The autonomous lawn mower and spectral SLAM. A.) 2 The ro )] (23) Optimization (last 10%) 0.68m 0.96m autonomous Batch lawn mower receiving 3,529 range measurements. This path minimizes the effect of (24) FastSLAM (last 10%) 0.73mmowing). 1.14m turns (a commonly used pattern for lawn Spectral SLAM recov ROP EKF traveled (last 10%) 0.65m the robot 1.3km receiving 1,816 range0.87m measurements. This path he standard 40 40same direction repeatedly. Again, spectral the SLAM Spectral SLAM (worst 10%) 1.17m 0.51m is able to recover t Spectral SLAM (best 10%) 0.28m 0.22m VI. C ONCLUSION 0.39m 0.35m Plaza 1Spectral SLAM (full path) ion experi- Fig. 3. Comparison of Range-Only SLAM Algorithms (Localization RMSE). 0 We proposed a novel formulation and solution for the rang 0 this paper. Boldface entries are best in column. í substantially from previo only SLAM problem that differs í th synthetic 20 í í 0 í í í 0 approaches. The via essence of thisfrom newmultispectral approach is to formul ved at each datasets were collected radio nodes ous lawn mower and spectral SLAM. A.) The robotic lawn mower platform. B.) In the first experiment, the robot travele Geoff Gordon UT Austin Feb, multivariate SLAM problem, which to deriv measurements. This path minimizes the as effecta offactorization heading error by balancing the number of leftallows turns with us an equal number

44 26 Structure from motion [Tomasi & Kanade, 1992]

45 Video textures Kalman filter works for some video textures steam grate example above fountain: observation = raw pixels (vector of reals over time) 27

46 28 Video textures, redux Original Kalman Filter PSR both models: 10 latent dimensions

47 28 Video textures, redux Original Kalman Filter PSR both models: 10 latent dimensions

48 29 Learning in the loop: option pricing pt pt 100 Price a financial derivative: psychic call holder gets to say I bought call 100 days ago underlying stock follows Black Scholes (unknown parameters)

49 30 Option pricing pt pt 100 Solution [Van Roy et al.]: use policy iteration 16 hand-picked features (e.g., poly history) initialize policy arbitrarily least-squares temporal differences (LSTD) to estimate value function policy := greedy; repeat

50 31 Option pricing Better solution: spectral SSID inside policy iteration 16 original features from Van Roy et al. 204 additional low-originality features e.g., linear fns of price history of underlying SSID picks best 16-d dynamics to explain feature evolution solve for value function in closed form

51 32 Policy iteration w/ spectral learning Expected payoff / $ of underlying Policy Iteration Iterations of PI Threshold LSTD (16) LSTD LARS-TD PSTD 0.82 /$ better than best competitor; 1.33 /$ better than best previously published data: 1,000,000 time steps

52 Avg. Prediction E Slot Ca Slot-car IMU data 6 Avg. Prediction Err. Slot Car A. 7 6 Hilbert 5 Space Embeddin IMU x 10 4 B. 8 HMM 7 3 Mean RR-HMM Last 6 2 LDS Embedded 5 IMU ï1 ï2 ï3 0 [Ko & Fox, 09; Boots et al., 10] Racetrack Predic Figure measureme Racetrack 4. Slot car inertial Prediction Horizon Figure 4. Slot car inertial measurement data. (A) The slot car platform and the IMU (top) and ar platform and the IMU (top) and the racetrack (bot(b) error Squared errorwith fordifferent predictio om).tom). (B) Squared for prediction esti1 0 mated models and Geoff Gordon UT Austin Feb, 2012baselines

53 Kalman < HSE-HMM < manifold Manuscript under review by AISTATS Manuscript under review by AISTATS Mean Obs. HMM (EM) HMM (E Mean Obs. Obs.HSE-HMM HSE-HM PreviousPrevious Obs. KalmanKalman Filter FilterManifold-HMM Manifold Gaussian RBFRBF Gaussian RMS Error RMS Error Different State Spaces Linear Linear C. 160 C Laplacian Eigenmap Laplacian Eigenmap Horizon 20Prediction RMS prediction errorhorizon (XY Prediction position) v. horizon comparison of kernels measurement unit (IMU). (A) The slot car platform: the car and IMU (top) 34

54 Algorithm intuition Use separate one-manifold learners to estimate structure in past, future result: noisy Gram matrices GX, GY signal is correlated between GX, GY noise is independent Look at eigensystem of GXGY suppresses noise, leaves signal Geoff Gordon MURI review Nov,

55 36 Summary General class of models for dynamical systems Fast, statistically consistent learning method Includes many well-known models & algorithms as special cases HMM, Kalman filter, n-gram, PSR, kernel versions SfM, subspace ID, Kalman video textures Also includes new models & algorithms that give state-of-the-art performance on interesting tasks range-only SLAM, PSR video textures, HSE-HMM

56 Papers B. Boots and G. An Online Spectral Learning Algorithm for Partially Observable Nonlinear Dynamical Systems. AAAI, B. Boots and G.% Predictive state temporal difference learning.%nips, B. Boots, S. M. Siddiqi, and G.% Closing the learning-planning loop with predictive state representations. RSS, L. Song, B. Boots, S. M. Siddiqi, G., and A. J. Smola.% Hilbert space embeddings of hidden Markov models.% ICML, (Best paper) 37

Learning about State. Geoff Gordon Machine Learning Department Carnegie Mellon University

Learning about State. Geoff Gordon Machine Learning Department Carnegie Mellon University Learning about State Geoff Gordon Machine Learning Department Carnegie Mellon University joint work with Byron Boots, Sajid Siddiqi, Le Song, Alex Smola 2 What s out there?...... ot-2 ot-1 ot ot+1 ot+2

More information

Learning about State. Geoff Gordon Machine Learning Department Carnegie Mellon University

Learning about State. Geoff Gordon Machine Learning Department Carnegie Mellon University Learning about State Geoff Gordon Machine Learning Department Carnegie Mellon University joint work with Byron Boots, Sajid Siddiqi, Le Song, Alex Smola What s out there?...... ot-2 ot-1 ot ot+1 ot+2 2

More information

An Online Spectral Learning Algorithm for Partially Observable Nonlinear Dynamical Systems

An Online Spectral Learning Algorithm for Partially Observable Nonlinear Dynamical Systems An Online Spectral Learning Algorithm for Partially Observable Nonlinear Dynamical Systems Byron Boots and Geoffrey J. Gordon AAAI 2011 Select Lab Carnegie Mellon University What is out there?...... o

More information

Reduced-Rank Hidden Markov Models

Reduced-Rank Hidden Markov Models Reduced-Rank Hidden Markov Models Sajid M. Siddiqi Byron Boots Geoffrey J. Gordon Carnegie Mellon University ... x 1 x 2 x 3 x τ y 1 y 2 y 3 y τ Sequence of observations: Y =[y 1 y 2 y 3... y τ ] Assume

More information

arxiv: v1 [cs.lg] 10 Jul 2012

arxiv: v1 [cs.lg] 10 Jul 2012 A Spectral Learning Approach to Range-Only SLAM arxiv:1207.2491v1 [cs.lg] 10 Jul 2012 Byron Boots Machine Learning Department Carnegie Mellon University Pittsburgh, PA 15213 beb@cs.cmu.edu Abstract Geoffrey

More information

A Constraint Generation Approach to Learning Stable Linear Dynamical Systems

A Constraint Generation Approach to Learning Stable Linear Dynamical Systems A Constraint Generation Approach to Learning Stable Linear Dynamical Systems Sajid M. Siddiqi Byron Boots Geoffrey J. Gordon Carnegie Mellon University NIPS 2007 poster W22 steam Application: Dynamic Textures

More information

Hilbert Space Embeddings of Hidden Markov Models

Hilbert Space Embeddings of Hidden Markov Models Hilbert Space Embeddings of Hidden Markov Models Le Song Carnegie Mellon University Joint work with Byron Boots, Sajid Siddiqi, Geoff Gordon and Alex Smola 1 Big Picture QuesJon Graphical Models! Dependent

More information

Spectral Learning Part I

Spectral Learning Part I Spectral Learning Part I Geoff Gordon http://www.cs.cmu.edu/~ggordon/ Machine Learning Department Carnegie Mellon University http://www.cs.cmu.edu/ ~ ggordon/spectral-learning/ What is spectral learning?

More information

Hilbert Space Embeddings of Predictive State Representations

Hilbert Space Embeddings of Predictive State Representations Hilbert Space Embeddings of Predictive State Representations Byron Boots Computer Science and Engineering Dept. University of Washington Seattle, WA Arthur Gretton Gatsby Unit University College London

More information

Hilbert Space Embeddings of Hidden Markov Models

Hilbert Space Embeddings of Hidden Markov Models Le Song lesong@cs.cmu.edu Byron Boots beb@cs.cmu.edu School of Computer Science, Carnegie Mellon University, Pittsburgh, PA 15213, USA Sajid M. Siddiqi siddiqi@google.com Google, Pittsburgh, PA 15213,

More information

Incremental Sparse GP Regression for Continuous-time Trajectory Estimation & Mapping

Incremental Sparse GP Regression for Continuous-time Trajectory Estimation & Mapping Incremental Sparse GP Regression for Continuous-time Trajectory Estimation & Mapping Xinyan Yan College of Computing Georgia Institute of Technology Atlanta, GA 30332, USA xinyan.yan@cc.gatech.edu Vadim

More information

Estimating Covariance Using Factorial Hidden Markov Models

Estimating Covariance Using Factorial Hidden Markov Models Estimating Covariance Using Factorial Hidden Markov Models João Sedoc 1,2 with: Jordan Rodu 3, Lyle Ungar 1, Dean Foster 1 and Jean Gallier 1 1 University of Pennsylvania Philadelphia, PA joao@cis.upenn.edu

More information

Two-Manifold Problems with Applications to Nonlinear System Identification

Two-Manifold Problems with Applications to Nonlinear System Identification Two-Manifold Problems with Applications to Nonlinear System Identification Byron Boots Geoffrey J. Gordon Machine Learning Department, Carnegie Mellon University, Pittsburgh, PA 1213 beb@cs.cmu.edu ggordon@cs.cmu.edu

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Bayes Filter Reminder. Kalman Filter Localization. Properties of Gaussians. Gaussians. Prediction. Correction. σ 2. Univariate. 1 2πσ e.

Bayes Filter Reminder. Kalman Filter Localization. Properties of Gaussians. Gaussians. Prediction. Correction. σ 2. Univariate. 1 2πσ e. Kalman Filter Localization Bayes Filter Reminder Prediction Correction Gaussians p(x) ~ N(µ,σ 2 ) : Properties of Gaussians Univariate p(x) = 1 1 2πσ e 2 (x µ) 2 σ 2 µ Univariate -σ σ Multivariate µ Multivariate

More information

Linear Regression (continued)

Linear Regression (continued) Linear Regression (continued) Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 6, 2017 1 / 39 Outline 1 Administration 2 Review of last lecture 3 Linear regression

More information

Mobile Robot Localization

Mobile Robot Localization Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations

More information

CS8803: Statistical Techniques in Robotics Byron Boots. Hilbert Space Embeddings

CS8803: Statistical Techniques in Robotics Byron Boots. Hilbert Space Embeddings CS8803: Statistical Techniques in Robotics Byron Boots Hilbert Space Embeddings 1 Motivation CS8803: STR Hilbert Space Embeddings 2 Overview Multinomial Distributions Marginal, Joint, Conditional Sum,

More information

EKF and SLAM. McGill COMP 765 Sept 18 th, 2017

EKF and SLAM. McGill COMP 765 Sept 18 th, 2017 EKF and SLAM McGill COMP 765 Sept 18 th, 2017 Outline News and information Instructions for paper presentations Continue on Kalman filter: EKF and extension to mapping Example of a real mapping system:

More information

Learning gradients: prescriptive models

Learning gradients: prescriptive models Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan

More information

Introduction to Graphical Models

Introduction to Graphical Models Introduction to Graphical Models The 15 th Winter School of Statistical Physics POSCO International Center & POSTECH, Pohang 2018. 1. 9 (Tue.) Yung-Kyun Noh GENERALIZATION FOR PREDICTION 2 Probabilistic

More information

Autonomous Navigation for Flying Robots

Autonomous Navigation for Flying Robots Computer Vision Group Prof. Daniel Cremers Autonomous Navigation for Flying Robots Lecture 6.2: Kalman Filter Jürgen Sturm Technische Universität München Motivation Bayes filter is a useful tool for state

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

arxiv: v1 [cs.lg] 29 Dec 2011

arxiv: v1 [cs.lg] 29 Dec 2011 Two-Manifold Problems arxiv:1112.6399v1 [cs.lg] 29 Dec 211 Byron Boots Machine Learning Department Carnegie Mellon University Pittsburgh, PA 15213 beb@cs.cmu.edu Abstract Recently, there has been much

More information

Predictive Timing Models

Predictive Timing Models Predictive Timing Models Pierre-Luc Bacon McGill University pbacon@cs.mcgill.ca Borja Balle McGill University bballe@cs.mcgill.ca Doina Precup McGill University dprecup@cs.mcgill.ca Abstract We consider

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Using the Kalman Filter for SLAM AIMS 2015

Using the Kalman Filter for SLAM AIMS 2015 Using the Kalman Filter for SLAM AIMS 2015 Contents Trivial Kinematics Rapid sweep over localisation and mapping (components of SLAM) Basic EKF Feature Based SLAM Feature types and representations Implementation

More information

Gaussian Processes as Continuous-time Trajectory Representations: Applications in SLAM and Motion Planning

Gaussian Processes as Continuous-time Trajectory Representations: Applications in SLAM and Motion Planning Gaussian Processes as Continuous-time Trajectory Representations: Applications in SLAM and Motion Planning Jing Dong jdong@gatech.edu 2017-06-20 License CC BY-NC-SA 3.0 Discrete time SLAM Downsides: Measurements

More information

Mathematical Formulation of Our Example

Mathematical Formulation of Our Example Mathematical Formulation of Our Example We define two binary random variables: open and, where is light on or light off. Our question is: What is? Computer Vision 1 Combining Evidence Suppose our robot

More information

Robotics. Mobile Robotics. Marc Toussaint U Stuttgart

Robotics. Mobile Robotics. Marc Toussaint U Stuttgart Robotics Mobile Robotics State estimation, Bayes filter, odometry, particle filter, Kalman filter, SLAM, joint Bayes filter, EKF SLAM, particle SLAM, graph-based SLAM Marc Toussaint U Stuttgart DARPA Grand

More information

L06. LINEAR KALMAN FILTERS. NA568 Mobile Robotics: Methods & Algorithms

L06. LINEAR KALMAN FILTERS. NA568 Mobile Robotics: Methods & Algorithms L06. LINEAR KALMAN FILTERS NA568 Mobile Robotics: Methods & Algorithms 2 PS2 is out! Landmark-based Localization: EKF, UKF, PF Today s Lecture Minimum Mean Square Error (MMSE) Linear Kalman Filter Gaussian

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

arxiv: v2 [stat.ml] 28 Feb 2018

arxiv: v2 [stat.ml] 28 Feb 2018 An Efficient, Expressive and Local Minima-free Method for Learning Controlled Dynamical Systems Ahmed Hefny Carnegie Mellon University Carlton Downey Carnegie Mellon University Geoffrey Gordon Carnegie

More information

Mobile Robot Localization

Mobile Robot Localization Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations

More information

L11. EKF SLAM: PART I. NA568 Mobile Robotics: Methods & Algorithms

L11. EKF SLAM: PART I. NA568 Mobile Robotics: Methods & Algorithms L11. EKF SLAM: PART I NA568 Mobile Robotics: Methods & Algorithms Today s Topic EKF Feature-Based SLAM State Representation Process / Observation Models Landmark Initialization Robot-Landmark Correlation

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms Saniv Kumar, Google Research, NY EECS-6898, Columbia University - Fall, 010 Saniv Kumar 9/13/010 EECS6898 Large Scale Machine Learning 1 Curse of Dimensionality Gaussian Mixture Models

More information

Human-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg

Human-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg Temporal Reasoning Kai Arras, University of Freiburg 1 Temporal Reasoning Contents Introduction Temporal Reasoning Hidden Markov Models Linear Dynamical Systems (LDS) Kalman Filter 2 Temporal Reasoning

More information

Final Exam, Machine Learning, Spring 2009

Final Exam, Machine Learning, Spring 2009 Name: Andrew ID: Final Exam, 10701 Machine Learning, Spring 2009 - The exam is open-book, open-notes, no electronics other than calculators. - The maximum possible score on this exam is 100. You have 3

More information

Machine Learning! in just a few minutes. Jan Peters Gerhard Neumann

Machine Learning! in just a few minutes. Jan Peters Gerhard Neumann Machine Learning! in just a few minutes Jan Peters Gerhard Neumann 1 Purpose of this Lecture Foundations of machine learning tools for robotics We focus on regression methods and general principles Often

More information

Hilbert Space Embeddings of Hidden Markov Models

Hilbert Space Embeddings of Hidden Markov Models Carnegie Mellon University Research Showcase @ CMU Machine Learning Department School of Computer Science 6-00 Hilbert Space Embeddings of Hidden Markov Models Le Song Carnegie Mellon University Byron

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Brief Introduction of Machine Learning Techniques for Content Analysis

Brief Introduction of Machine Learning Techniques for Content Analysis 1 Brief Introduction of Machine Learning Techniques for Content Analysis Wei-Ta Chu 2008/11/20 Outline 2 Overview Gaussian Mixture Model (GMM) Hidden Markov Model (HMM) Support Vector Machine (SVM) Overview

More information

Hidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010

Hidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010 Hidden Markov Models Aarti Singh Slides courtesy: Eric Xing Machine Learning 10-701/15-781 Nov 8, 2010 i.i.d to sequential data So far we assumed independent, identically distributed data Sequential data

More information

Particle Filters; Simultaneous Localization and Mapping (Intelligent Autonomous Robotics) Subramanian Ramamoorthy School of Informatics

Particle Filters; Simultaneous Localization and Mapping (Intelligent Autonomous Robotics) Subramanian Ramamoorthy School of Informatics Particle Filters; Simultaneous Localization and Mapping (Intelligent Autonomous Robotics) Subramanian Ramamoorthy School of Informatics Recap: State Estimation using Kalman Filter Project state and error

More information

SLAM Techniques and Algorithms. Jack Collier. Canada. Recherche et développement pour la défense Canada. Defence Research and Development Canada

SLAM Techniques and Algorithms. Jack Collier. Canada. Recherche et développement pour la défense Canada. Defence Research and Development Canada SLAM Techniques and Algorithms Jack Collier Defence Research and Development Canada Recherche et développement pour la défense Canada Canada Goals What will we learn Gain an appreciation for what SLAM

More information

Techniques for Dimensionality Reduction. PCA and Other Matrix Factorization Methods

Techniques for Dimensionality Reduction. PCA and Other Matrix Factorization Methods Techniques for Dimensionality Reduction PCA and Other Matrix Factorization Methods Outline Principle Compoments Analysis (PCA) Example (Bishop, ch 12) PCA as a mixture model variant With a continuous latent

More information

From Bayes to Extended Kalman Filter

From Bayes to Extended Kalman Filter From Bayes to Extended Kalman Filter Michal Reinštein Czech Technical University in Prague Faculty of Electrical Engineering, Department of Cybernetics Center for Machine Perception http://cmp.felk.cvut.cz/

More information

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein Kalman filtering and friends: Inference in time series models Herke van Hoof slides mostly by Michael Rubinstein Problem overview Goal Estimate most probable state at time k using measurement up to time

More information

Final Exam December 12, 2017

Final Exam December 12, 2017 Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes

More information

STA 414/2104: Machine Learning

STA 414/2104: Machine Learning STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far

More information

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010 Linear Regression Aarti Singh Machine Learning 10-701/15-781 Sept 27, 2010 Discrete to Continuous Labels Classification Sports Science News Anemic cell Healthy cell Regression X = Document Y = Topic X

More information

Dual Estimation and the Unscented Transformation

Dual Estimation and the Unscented Transformation Dual Estimation and the Unscented Transformation Eric A. Wan ericwan@ece.ogi.edu Rudolph van der Merwe rudmerwe@ece.ogi.edu Alex T. Nelson atnelson@ece.ogi.edu Oregon Graduate Institute of Science & Technology

More information

Introduction to Mobile Robotics SLAM: Simultaneous Localization and Mapping

Introduction to Mobile Robotics SLAM: Simultaneous Localization and Mapping Introduction to Mobile Robotics SLAM: Simultaneous Localization and Mapping Wolfram Burgard, Cyrill Stachniss, Kai Arras, Maren Bennewitz What is SLAM? Estimate the pose of a robot and the map of the environment

More information

Kernel Bayes Rule: Nonparametric Bayesian inference with kernels

Kernel Bayes Rule: Nonparametric Bayesian inference with kernels Kernel Bayes Rule: Nonparametric Bayesian inference with kernels Kenji Fukumizu The Institute of Statistical Mathematics NIPS 2012 Workshop Confluence between Kernel Methods and Graphical Models December

More information

Hidden Markov models

Hidden Markov models Hidden Markov models Charles Elkan November 26, 2012 Important: These lecture notes are based on notes written by Lawrence Saul. Also, these typeset notes lack illustrations. See the classroom lectures

More information

Linear Dynamical Systems

Linear Dynamical Systems Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations

More information

Machine Learning. CUNY Graduate Center, Spring Lectures 11-12: Unsupervised Learning 1. Professor Liang Huang.

Machine Learning. CUNY Graduate Center, Spring Lectures 11-12: Unsupervised Learning 1. Professor Liang Huang. Machine Learning CUNY Graduate Center, Spring 2013 Lectures 11-12: Unsupervised Learning 1 (Clustering: k-means, EM, mixture models) Professor Liang Huang huang@cs.qc.cuny.edu http://acl.cs.qc.edu/~lhuang/teaching/machine-learning

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: ony Jebara Kalman Filtering Linear Dynamical Systems and Kalman Filtering Structure from Motion Linear Dynamical Systems Audio: x=pitch y=acoustic waveform Vision: x=object

More information

Inference Machines for Nonparametric Filter Learning

Inference Machines for Nonparametric Filter Learning Inference Machines for Nonparametric Filter Learning Arun Venkatraman, Wen Sun, Martial Hebert, Byron Boots 2, J. Andrew Bagnell Robotics Institute, Carnegie Mellon University, Pittsburgh, PA 523 2 School

More information

CSCI567 Machine Learning (Fall 2014)

CSCI567 Machine Learning (Fall 2014) CSCI567 Machine Learning (Fall 24) Drs. Sha & Liu {feisha,yanliu.cs}@usc.edu October 2, 24 Drs. Sha & Liu ({feisha,yanliu.cs}@usc.edu) CSCI567 Machine Learning (Fall 24) October 2, 24 / 24 Outline Review

More information

Dimensionality Reduction by Unsupervised Regression

Dimensionality Reduction by Unsupervised Regression Dimensionality Reduction by Unsupervised Regression Miguel Á. Carreira-Perpiñán, EECS, UC Merced http://faculty.ucmerced.edu/mcarreira-perpinan Zhengdong Lu, CSEE, OGI http://www.csee.ogi.edu/~zhengdon

More information

CS Homework 3. October 15, 2009

CS Homework 3. October 15, 2009 CS 294 - Homework 3 October 15, 2009 If you have questions, contact Alexandre Bouchard (bouchard@cs.berkeley.edu) for part 1 and Alex Simma (asimma@eecs.berkeley.edu) for part 2. Also check the class website

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Advanced Introduction to Machine Learning

Advanced Introduction to Machine Learning 10-715 Advanced Introduction to Machine Learning Homework Due Oct 15, 10.30 am Rules Please follow these guidelines. Failure to do so, will result in loss of credit. 1. Homework is due on the due date

More information

Spectral Probabilistic Modeling and Applications to Natural Language Processing

Spectral Probabilistic Modeling and Applications to Natural Language Processing School of Computer Science Spectral Probabilistic Modeling and Applications to Natural Language Processing Doctoral Thesis Proposal Ankur Parikh Thesis Committee: Eric Xing, Geoff Gordon, Noah Smith, Le

More information

Latent Tree Approximation in Linear Model

Latent Tree Approximation in Linear Model Latent Tree Approximation in Linear Model Navid Tafaghodi Khajavi Dept. of Electrical Engineering, University of Hawaii, Honolulu, HI 96822 Email: navidt@hawaii.edu ariv:1710.01838v1 [cs.it] 5 Oct 2017

More information

UAV Navigation: Airborne Inertial SLAM

UAV Navigation: Airborne Inertial SLAM Introduction UAV Navigation: Airborne Inertial SLAM Jonghyuk Kim Faculty of Engineering and Information Technology Australian National University, Australia Salah Sukkarieh ARC Centre of Excellence in

More information

Simultaneous Localization and Mapping

Simultaneous Localization and Mapping Simultaneous Localization and Mapping Miroslav Kulich Intelligent and Mobile Robotics Group Czech Institute of Informatics, Robotics and Cybernetics Czech Technical University in Prague Winter semester

More information

Final Exam December 12, 2017

Final Exam December 12, 2017 Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes

More information

Analysis of Spectral Kernel Design based Semi-supervised Learning

Analysis of Spectral Kernel Design based Semi-supervised Learning Analysis of Spectral Kernel Design based Semi-supervised Learning Tong Zhang IBM T. J. Watson Research Center Yorktown Heights, NY 10598 Rie Kubota Ando IBM T. J. Watson Research Center Yorktown Heights,

More information

ROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino

ROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino ROBOTICS 01PEEQW Basilio Bona DAUIN Politecnico di Torino Probabilistic Fundamentals in Robotics Gaussian Filters Course Outline Basic mathematical framework Probabilistic models of mobile robots Mobile

More information

arxiv: v1 [cs.lg] 12 Dec 2009

arxiv: v1 [cs.lg] 12 Dec 2009 Closing the Learning-Planning Loop with Predictive State Representations Byron Boots Machine Learning Department Carnegie Mellon University Pittsburgh, PA 15213 beb@cs.cmu.edu Sajid M. Siddiqi Robotics

More information

Restricted Boltzmann Machines for Collaborative Filtering

Restricted Boltzmann Machines for Collaborative Filtering Restricted Boltzmann Machines for Collaborative Filtering Authors: Ruslan Salakhutdinov Andriy Mnih Geoffrey Hinton Benjamin Schwehn Presentation by: Ioan Stanculescu 1 Overview The Netflix prize problem

More information

PROBABILISTIC REASONING OVER TIME

PROBABILISTIC REASONING OVER TIME PROBABILISTIC REASONING OVER TIME In which we try to interpret the present, understand the past, and perhaps predict the future, even when very little is crystal clear. Outline Time and uncertainty Inference:

More information

Chapter 3. Tohru Katayama

Chapter 3. Tohru Katayama Subspace Methods for System Identification Chapter 3 Tohru Katayama Subspace Methods Reading Group UofA, Edmonton Barnabás Póczos May 14, 2009 Preliminaries before Linear Dynamical Systems Hidden Markov

More information

CS 532: 3D Computer Vision 6 th Set of Notes

CS 532: 3D Computer Vision 6 th Set of Notes 1 CS 532: 3D Computer Vision 6 th Set of Notes Instructor: Philippos Mordohai Webpage: www.cs.stevens.edu/~mordohai E-mail: Philippos.Mordohai@stevens.edu Office: Lieb 215 Lecture Outline Intro to Covariance

More information

CSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression

CSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression CSC2515 Winter 2015 Introduction to Machine Learning Lecture 2: Linear regression All lecture slides will be available as.pdf on the course website: http://www.cs.toronto.edu/~urtasun/courses/csc2515/csc2515_winter15.html

More information

TSRT14: Sensor Fusion Lecture 9

TSRT14: Sensor Fusion Lecture 9 TSRT14: Sensor Fusion Lecture 9 Simultaneous localization and mapping (SLAM) Gustaf Hendeby gustaf.hendeby@liu.se TSRT14 Lecture 9 Gustaf Hendeby Spring 2018 1 / 28 Le 9: simultaneous localization and

More information

CS Machine Learning Qualifying Exam

CS Machine Learning Qualifying Exam CS Machine Learning Qualifying Exam Georgia Institute of Technology March 30, 2017 The exam is divided into four areas: Core, Statistical Methods and Models, Learning Theory, and Decision Processes. There

More information

Laplacian Agent Learning: Representation Policy Iteration

Laplacian Agent Learning: Representation Policy Iteration Laplacian Agent Learning: Representation Policy Iteration Sridhar Mahadevan Example of a Markov Decision Process a1: $0 Heaven $1 Earth What should the agent do? a2: $100 Hell $-1 V a1 ( Earth ) = f(0,1,1,1,1,...)

More information

Learning SVM Classifiers with Indefinite Kernels

Learning SVM Classifiers with Indefinite Kernels Learning SVM Classifiers with Indefinite Kernels Suicheng Gu and Yuhong Guo Dept. of Computer and Information Sciences Temple University Support Vector Machines (SVMs) (Kernel) SVMs are widely used in

More information

Overview of Statistical Tools. Statistical Inference. Bayesian Framework. Modeling. Very simple case. Things are usually more complicated

Overview of Statistical Tools. Statistical Inference. Bayesian Framework. Modeling. Very simple case. Things are usually more complicated Fall 3 Computer Vision Overview of Statistical Tools Statistical Inference Haibin Ling Observation inference Decision Prior knowledge http://www.dabi.temple.edu/~hbling/teaching/3f_5543/index.html Bayesian

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 12 Dynamical Models CS/CNS/EE 155 Andreas Krause Homework 3 out tonight Start early!! Announcements Project milestones due today Please email to TAs 2 Parameter learning

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Hidden Markov Models Barnabás Póczos & Aarti Singh Slides courtesy: Eric Xing i.i.d to sequential data So far we assumed independent, identically distributed

More information

Dimension Reduction. David M. Blei. April 23, 2012

Dimension Reduction. David M. Blei. April 23, 2012 Dimension Reduction David M. Blei April 23, 2012 1 Basic idea Goal: Compute a reduced representation of data from p -dimensional to q-dimensional, where q < p. x 1,...,x p z 1,...,z q (1) We want to do

More information

STA 4273H: Sta-s-cal Machine Learning

STA 4273H: Sta-s-cal Machine Learning STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our

More information

November 28 th, Carlos Guestrin 1. Lower dimensional projections

November 28 th, Carlos Guestrin 1. Lower dimensional projections PCA Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University November 28 th, 2007 1 Lower dimensional projections Rather than picking a subset of the features, we can new features that are

More information

Localization. Howie Choset Adapted from slides by Humphrey Hu, Trevor Decker, and Brad Neuman

Localization. Howie Choset Adapted from slides by Humphrey Hu, Trevor Decker, and Brad Neuman Localization Howie Choset Adapted from slides by Humphrey Hu, Trevor Decker, and Brad Neuman Localization General robotic task Where am I? Techniques generalize to many estimation tasks System parameter

More information

Expectation Propagation in Dynamical Systems

Expectation Propagation in Dynamical Systems Expectation Propagation in Dynamical Systems Marc Peter Deisenroth Joint Work with Shakir Mohamed (UBC) August 10, 2012 Marc Deisenroth (TU Darmstadt) EP in Dynamical Systems 1 Motivation Figure : Complex

More information

26 : Spectral GMs. Lecturer: Eric P. Xing Scribes: Guillermo A Cidre, Abelino Jimenez G.

26 : Spectral GMs. Lecturer: Eric P. Xing Scribes: Guillermo A Cidre, Abelino Jimenez G. 10-708: Probabilistic Graphical Models, Spring 2015 26 : Spectral GMs Lecturer: Eric P. Xing Scribes: Guillermo A Cidre, Abelino Jimenez G. 1 Introduction A common task in machine learning is to work with

More information

Learning Predictive Models of a Depth Camera & Manipulator from Raw Execution Traces

Learning Predictive Models of a Depth Camera & Manipulator from Raw Execution Traces Learning Predictive Models of a Depth Camera & Manipulator from Raw Execution Traces Byron Boots Arunkumar Byravan Dieter Fox Abstract In this paper, we attack the problem of learning a predictive model

More information

Review: Probabilistic Matrix Factorization. Probabilistic Matrix Factorization (PMF)

Review: Probabilistic Matrix Factorization. Probabilistic Matrix Factorization (PMF) Case Study 4: Collaborative Filtering Review: Probabilistic Matrix Factorization Machine Learning for Big Data CSE547/STAT548, University of Washington Emily Fox February 2 th, 214 Emily Fox 214 1 Probabilistic

More information

Kernel-Based Contrast Functions for Sufficient Dimension Reduction

Kernel-Based Contrast Functions for Sufficient Dimension Reduction Kernel-Based Contrast Functions for Sufficient Dimension Reduction Michael I. Jordan Departments of Statistics and EECS University of California, Berkeley Joint work with Kenji Fukumizu and Francis Bach

More information

CS491/691: Introduction to Aerial Robotics

CS491/691: Introduction to Aerial Robotics CS491/691: Introduction to Aerial Robotics Topic: State Estimation Dr. Kostas Alexis (CSE) World state (or system state) Belief state: Our belief/estimate of the world state World state: Real state of

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,

More information

Dynamic Data Modeling, Recognition, and Synthesis. Rui Zhao Thesis Defense Advisor: Professor Qiang Ji

Dynamic Data Modeling, Recognition, and Synthesis. Rui Zhao Thesis Defense Advisor: Professor Qiang Ji Dynamic Data Modeling, Recognition, and Synthesis Rui Zhao Thesis Defense Advisor: Professor Qiang Ji Contents Introduction Related Work Dynamic Data Modeling & Analysis Temporal localization Insufficient

More information

Modeling Multiple-mode Systems with Predictive State Representations

Modeling Multiple-mode Systems with Predictive State Representations Modeling Multiple-mode Systems with Predictive State Representations Britton Wolfe Computer Science Indiana University-Purdue University Fort Wayne wolfeb@ipfw.edu Michael R. James AI and Robotics Group

More information

1 Bayesian Linear Regression (BLR)

1 Bayesian Linear Regression (BLR) Statistical Techniques in Robotics (STR, S15) Lecture#10 (Wednesday, February 11) Lecturer: Byron Boots Gaussian Properties, Bayesian Linear Regression 1 Bayesian Linear Regression (BLR) In linear regression,

More information

Rao-Blackwellized Particle Filtering for 6-DOF Estimation of Attitude and Position via GPS and Inertial Sensors

Rao-Blackwellized Particle Filtering for 6-DOF Estimation of Attitude and Position via GPS and Inertial Sensors Rao-Blackwellized Particle Filtering for 6-DOF Estimation of Attitude and Position via GPS and Inertial Sensors GRASP Laboratory University of Pennsylvania June 6, 06 Outline Motivation Motivation 3 Problem

More information