Probabilistic Robotics, Sebastian Thrun, 2005 Page 36, 37. <1 x = nf ) =1/ 3 <1 x = f ) =1 p(x = f ) = 0.01 p(x = nf ) = 0.

Size: px
Start display at page:

Download "Probabilistic Robotics, Sebastian Thrun, 2005 Page 36, 37. <1 x = nf ) =1/ 3 <1 x = f ) =1 p(x = f ) = 0.01 p(x = nf ) = 0."

Transcription

1 Probabilistic Robotics, Sebastian Thrun, 005 Page 36, 37 Method : Denote: z n : the sensor measurement in each time x: the state of the sensor f: the state of the sensor is faulty nf: the state of the sensor is not faulty We know: p(z < x nf ) / 3 p(z < x f ) p(x f ) 0.0 p(x nf ) 0.99 when N p(x z ) p(z x) p(x) p(z ) x' f,nf p(z x) p(x) p(z x') p(x') *0.0 p(x f z <) * *0.99 when N pz (, z xpx ) ( ) pz (, z xpx ) ( ) p(x z,z ) pz (, z) pz (, z x') ') x' f, nf *0.0 f z <, z < ) *0.0 + ( ) *0.99 3

2 when N0 pz ( :0 xpx ) ( ) pz ( :0 xpx ) ( ) p(x z :0) pz ( ) pz ( x') ') :0 :0 x' f, nf *0.0 f z:0 < ) *0.0 + ( ) * When Nn Method : Denote: X : the sensor is faulty X : the sensor is right Y: the event that all the N outputs are below m. px ( ) 0.0 PX ( ) 0.99 py ( X) py ( X) ( ) 3 px ( Y) N px (, Y) py ( X) PX ( ) py ( ) PY ( X) P( X) + PY ( X ) P( X ) 0.0 N ( ) * N,,, 0. The posterior probability of a sensor fault is 0.094, , 0.43, ,0.705,0.8804,0.9567,0.985,0.9950,0.9983

3 (a) Suppose Day is a sunny day. What is the probability of the following sequence of days: Day cloudy, Day3 cloudy, Day4 rainy? Denote: s: sunny, c: cloudy, r: rainy state value x : weather of day n by Markov assumption : n, x, x, x) x, x, x), x, x), x, x x) x), x x) ) ), x x) , x, x) x, x), x) 3 x) x) ) ) 3 3 x) pc ( s) 0. x) pcc ( ) 0.4 x) pr ( c) 0. Thus p( x, x, x x ) p( x x ) p( x x ) p( x x ) (b) Write a simulator that can randomly generate sequences of weathers" from this state transition function. [Reference code]: sequencedays 0; s ; c 0; r 0; % the possible value of state T [8 0; 4 4 ; 6 ]'/0; % Markov state transition matrix xp [s c r]'; % the state of the previous day weather {'sunny ', 'cloudy', 'rainy '}; for j:sequencedays xt T * xp; xn [0 0 0]'; % the state of the next day randomprob rand(); cumulativeprob 0; for i:3 cumulativeprob cumulativeprob + xt(i); if (randomprob < cumulativeprob) xn(i) ; break;

4 for i:3 if (xn(i) ) disp(weather{i}); break; [Output (may not exactly the same value)]: sunny, cloudy, sunny, cloudy, sunny, sunny, cloudy, cloudy, sunny, sunny, (c) Use your simulator to determine the stationary distribution of this Markov chain. The stationary distribution measures the probability that a random day will be sunny, cloudy, or rainy. Stationary distribution can be approximated by the frequency of weather far into the future. [Reference code]: E.g. set the statistic sequence days as: sequencedays ; Add the count for statistic in the code: if (xn(i) ) disp(weather{i}); weathercnt(i) weathercnt(i) + ; break; [Output (may not exactly the same value)]: sunny has 6459 days in days, prob cloudy has 8603 days in days, prob rainy has 7439 days in days, prob Thus the stationary distribution of this Markov chain is [0.645, 0.860, 0.074] T. (d) Can you devise a closed-form solution to calculating the stationary distribution based on the state transition matrix above? The stationary distribution means that λ x T x (λ ), in this case, x is the linearly normalized eigen vector of matrix T, when eigen value λ. [Reference code]: [V, D] eig(t) xbar V(:, ) % this st column of V corresponding to eigen value xbar xbar / sum(xbar) [Output]: xbar (e) What is the entropy of the stationary distribution? For stationary Markov chain {Xi} with stationary distribution X and transition matrix P. The entropy is given by: n H( X) lim H( X,..., Xn) lim H( Xi Xi,..., X), by chain rule. n n n n i n p( X ) n n i i i lim H( X X ) p( X )log n

5 [Reference code]: Hx (-) * sum(xbar.* log(xbar)); [Output]: Hx.98 (f) Using Bayes rule, compute the probability table of yesterday's weather given today's weather. (It is okay to provide the probabilities numerically, and it is also okay to rely on results from previous questions in this exercise.) Bayes rule and total probability: i xi ) i ) i xi ) i ) i xi) ) x ) ) i i i i xi If we use stationary probability as the prior: ) [0.649, 0.857, 0.074] T i [Reference code]: Tr zeros(3); for i:3 % weather state: today % weather state: yesterday % calculate p(x_t_ j x_t i) Tr(i, j) T(i, j) * xbar(j) / sum(t(i, :)'.* xbar); for j:3 [Output]: Tr yesterday s weather Today s weather sunny cloudy rainy sunny cloudy rainy If we use uniform probability as the prior: T i ) [,, ] [Reference code]: xbar [ ]' / 3; Tr zeros(3); for i:3 % weather state: today for j:3 % weather state: yesterday % calculate p(x_t_ j x_t i) Tr(i, j) T(i, j) * xbar(j) / sum(t(i, :)'.* xbar); [Output]: Tr

6 Today s weather Yesterday s weather sunny cloudy rainy sunny cloudy rainy (g) Suppose we added seasons to our model. The state transition function above would only apply to the Summer, whereas different ones would apply to Winter, Spring, and Fall. Would this violate the Markov property of this process? Explain your answer. The system still satisfies Markov assumption. It can be viewed as a time-varying Markov Chain (though time-varying, the value of x t+ is still determined by x t only). Alternatively, add season s t as an additional state variable of four values. By ex-ting the weather transition matrix to x, we can still account for all possible transitions from {x t, s t } to {x t+, s t+ }:, s x, s ), s x, s) t+ t+ : t : t t+ t+ t t

7 (a) Suppose Day is sunny (this is known for a fact), and in the subsequent four days our sensor observes cloudy, cloudy, rainy, sunny. What is the probability that Day 5 is indeed sunny as predicted by our sensor? Method : Denote: s: sunny, c: cloudy, r: rainy state value: x : weather of day n, we know x s, x r n 4 sensor data: z n: weather of day n, we know z ~5 ccrs,,, By applying Bayes' rule and Markov assumption: Bel( x ) p( x x, z, z, z, z ) pz ( 5 x5), z, z3, z4, x5) pz ( 5 x5) 5 x, z, z3, z4), z, z, z, z) pz ( x, z, z, z) pz ( x) x, z, z, z) pz ( 5 x5) 5 x, z, z3, z4) x5 s, c, r As we can see in the sensor measurement transition matrix, z 4 rainy, means x 4 rainy ' ' we have x, z, z, z) x r) x) x r) Bel(x 5 s) ' x5 s, c, r p(z s x s) p(x s x r) " % $ p(z 5 s x ' ' 5 ) p(x 5 x 4 r) ' $ ' # x ' 5 s,c,r & from the previous question, we know the transition matrix 0.6*0. 0.6* * *

8 Method : Note that when the weather is rainy, the sensor reports that with absolute certainty. Therefore, we can just analyze the transition from day 4 to day5. We are looking to find p(x 5 s z 5 s) when x 4 r : p(x 5 s z 5 s) p(z 5 s x 5 s) p(x 5 s x 4 r) p(z 5 s) p(z s x s) p(x s x r) " % $ p(z 5 s x ' ' 5 ) p(x 5 x 4 r) ' $ ' # x ' 5 s,c,r & from the previous question, we know the transition matrix 0.6*0. 0.6* * * (b) Once again, suppose Day is known to be sunny. At Days through 4, the sensor measures sunny, sunny, rainy. For each of the Days through 4, what is the most likely weather on that day? Answer the question in two ways: one in which only the data available to the day in question is used, and one in hindsight, where data from future days is also available. For day : For the st case we only know the measures until this day: Method : Bel( x ) p( x x, z ) η pz ( x, x) x) η pz ( x) x) η η /9 / 9, normalizate the sum of prob. to 0 Method :

9 we already know x z s ' x s, c, r s Also the state & measurement transition matrices. Bel( x ) p( x x, z ) pz ( x) x) ' ' pz ( x) x) pz ( x) x) pz ( x) Bel(x s) (0.6*0.8)/(0.6* *0.+0*0) Bel(x c) (0.3*0.)/(0.6* *0.+0*0) 0. Bel(x r) 0 So its most likely weather is sunny with probability 88.89%

10 For day 3: For the st case we only know the measures until this day: Method : x, z ) 3 :3 η pz ( x, x, z) x, z) η pz ( x), x x, z) x η pz ( x) x) x, z) x 8 η pz ( 3 x3) 3 x) x ηpz ( 3 x3).0 η So its most likely weather is sunny with probability 87.8% For day 4: in the case we know the measures of all those days: x, z ) x, z) 4 : So its most likely weather is rainy with probability 00%

11 If we KNOW future measures. For day : For the nd case we know the measures of all those days: we already know x s, x r z s, z s, z r Also the state & measurement transition matrices. Bel( x ) p( x x, z, x ) p( x x, z ), since we already know z, as well as x :4 4 :4 4 4, x, z ) pz ( :4 : 4 :4 :4 :4 :4 3 4 :4 :4 :4 :4 x, x ) p( x, x ) x) pz ( :4 x) pz ( x) pz (, x) pz ( x), z, z) x) x) pz ( x) ) pz ( x) ) x) pz ( x) pz ( 3:4 x) pz ( x) x) pz ( x) 3, z3:4 x), Total prob. pz ( x) η x) pz ( x) x) pz ( x), Bayes rule x3 3 3:4 3 x3 η x) pz ( x) x) pz ( x) pz ( x, z), Bayes rule x3 η x) pz ( x) x) pz ( x) pz ( x), Bayes rule x3 η x) pz ( x) x) pz ( x), z x), Total Prob x3 x4 η x) pz ( x) x) pz ( x) x) pz ( x, x), Bayes rule x3 x4 η x) pz ( x) x) pz ( x) x) pz ( x), Markov assumption x3 x4

12 η x) pz ( x) x) pz ( x) x) pz ( x), Markov assumption x3 x4 assume m x) pz ( x) x when x s, m p( x r x s) p( z r x r) when x c, m p( x r x c) p( z r x r) when x r, m p( x r x r) p( z r x r) η x) pz ( x) x) pz ( x) m x3 assume n x) pz ( x) m x when x s, n p( x s x s) p( z s x s)* c x spz ) ( s x c)* r x spz ) ( s x r)* 0. 0.*0.3* when x c, n p( x s x c) p( z s x s)* c x cp ) ( z3 s x3 c)*0. + r x cpz ) ( s x r)*0. 0.4*0.3* when x r, n p( x s x r) p( z s x s)* c x rpz ) ( s x c)* r x rpz ) ( s x r)*0. 0.6*0.3* η x spz ) ( s x) η η So its most likely weather is sunny with probability 80%

13 For day 3: For the nd case we know the measures of all those days: we already know x s, z s, z s, z r,also we can deduce x r Bel( x ) p( x x, z ) p( x, x, z ) 3 3 :4 3 :4, z:4) pz ( 4 x3) 3, z:3, x) :4 pz ( 4 x3) 3, z:3, x) :4 pz ( 4 x3) pz ( 3 x3, z, x) p( x3, z, x) :4 pz ( x) pz ( x), x, z, x) :4 :4 :4 : x pz ( x) pz ( x) x), z, x) x pz ( x) pz ( x) x) pz ( x), x) x ) 4 x3) pz ( 3 x3) 3 x) pz ( x) x) assume m x) pz ( x) x) x x 3 when x s, m p( x s x s) p( z s x s) p( x s x s) s x cpz ) ( s x cpx ) ( c x s) 3 + s x rpz ) ( s x rpx ) ( r x s) 3 0.8*0.6* *0.3* when x c, m p( x c x s) p( z s x s) p( x s x s) c x cpz ) ( s x cpx ) ( c x s) + c x rpz ) ( s x rpx ) ( r x s) 3 0.*0.6* *0.3* when x r, m p( x r x s) p( z s x s) p( x s x s) r x cpz ) ( s x cpx ) ( c x s) 3 + r x rpz ) ( s x rpx ) ( r x s) *0.3* η 4 r x3) pz ( 3 s x3) 0.00 η η So its most likely weather is cloudy with probability 00%

14 (c) Consider the same situation (Day is sunny, the measurements for Days, 3, and 4 are sunny, sunny, rainy). What is the most likely sequence of weather for Days through 4? What is the probability of this most likely sequence? The probability of the sequence of weather is given by: x, z ) :4 :4 pz ( :4 x, x:4 ), x:4) :4 ) pz ( :4 x, x:4 ) :4 x ) :4 η pz ( 4 x4) pz ( 3 x3) pz ( x)* 4 x3) 3 x) x) Using the sensor model and transition matrix to calculate it, we can get the most likely sequence of weather is s, c, r, which has / ( ) 80% of occurring. There is a 0% probability of c, c, r, and 0% probability of all other possibles.

Mobile Robotics II: Simultaneous localization and mapping

Mobile Robotics II: Simultaneous localization and mapping Mobile Robotics II: Simultaneous localization and mapping Introduction: probability theory, estimation Miroslav Kulich Intelligent and Mobile Robotics Group Gerstner Laboratory for Intelligent Decision

More information

CS325 Artificial Intelligence Ch. 15,20 Hidden Markov Models and Particle Filtering

CS325 Artificial Intelligence Ch. 15,20 Hidden Markov Models and Particle Filtering CS325 Artificial Intelligence Ch. 15,20 Hidden Markov Models and Particle Filtering Cengiz Günay, Emory Univ. Günay Ch. 15,20 Hidden Markov Models and Particle FilteringSpring 2013 1 / 21 Get Rich Fast!

More information

Markov localization uses an explicit, discrete representation for the probability of all position in the state space.

Markov localization uses an explicit, discrete representation for the probability of all position in the state space. Markov Kalman Filter Localization Markov localization localization starting from any unknown position recovers from ambiguous situation. However, to update the probability of all positions within the whole

More information

Introduction to Mobile Robotics Probabilistic Robotics

Introduction to Mobile Robotics Probabilistic Robotics Introduction to Mobile Robotics Probabilistic Robotics Wolfram Burgard 1 Probabilistic Robotics Key idea: Explicit representation of uncertainty (using the calculus of probability theory) Perception Action

More information

Modeling and state estimation Examples State estimation Probabilities Bayes filter Particle filter. Modeling. CSC752 Autonomous Robotic Systems

Modeling and state estimation Examples State estimation Probabilities Bayes filter Particle filter. Modeling. CSC752 Autonomous Robotic Systems Modeling CSC752 Autonomous Robotic Systems Ubbo Visser Department of Computer Science University of Miami February 21, 2017 Outline 1 Modeling and state estimation 2 Examples 3 State estimation 4 Probabilities

More information

Probabilistic Robotics

Probabilistic Robotics Probabilistic Robotics Overview of probability, Representing uncertainty Propagation of uncertainty, Bayes Rule Application to Localization and Mapping Slides from Autonomous Robots (Siegwart and Nourbaksh),

More information

Lecture 6: Entropy Rate

Lecture 6: Entropy Rate Lecture 6: Entropy Rate Entropy rate H(X) Random walk on graph Dr. Yao Xie, ECE587, Information Theory, Duke University Coin tossing versus poker Toss a fair coin and see and sequence Head, Tail, Tail,

More information

Robots Autónomos. Depto. CCIA. 2. Bayesian Estimation and sensor models. Domingo Gallardo

Robots Autónomos. Depto. CCIA. 2. Bayesian Estimation and sensor models.  Domingo Gallardo Robots Autónomos 2. Bayesian Estimation and sensor models Domingo Gallardo Depto. CCIA http://www.rvg.ua.es/master/robots References Recursive State Estimation: Thrun, chapter 2 Sensor models and robot

More information

Markov Chains and MCMC

Markov Chains and MCMC Markov Chains and MCMC CompSci 590.02 Instructor: AshwinMachanavajjhala Lecture 4 : 590.02 Spring 13 1 Recap: Monte Carlo Method If U is a universe of items, and G is a subset satisfying some property,

More information

PROBABILITY AND INFERENCE

PROBABILITY AND INFERENCE PROBABILITY AND INFERENCE Progress Report We ve finished Part I: Problem Solving! Part II: Reasoning with uncertainty Part III: Machine Learning 1 Today Random variables and probabilities Joint, marginal,

More information

18.440: Lecture 33 Markov Chains

18.440: Lecture 33 Markov Chains 18.440: Lecture 33 Markov Chains Scott Sheffield MIT 1 Outline Markov chains Examples Ergodicity and stationarity 2 Outline Markov chains Examples Ergodicity and stationarity 3 Markov chains Consider a

More information

18.600: Lecture 32 Markov Chains

18.600: Lecture 32 Markov Chains 18.600: Lecture 32 Markov Chains Scott Sheffield MIT Outline Markov chains Examples Ergodicity and stationarity Outline Markov chains Examples Ergodicity and stationarity Markov chains Consider a sequence

More information

Plan for today. ! Part 1: (Hidden) Markov models. ! Part 2: String matching and read mapping

Plan for today. ! Part 1: (Hidden) Markov models. ! Part 2: String matching and read mapping Plan for today! Part 1: (Hidden) Markov models! Part 2: String matching and read mapping! 2.1 Exact algorithms! 2.2 Heuristic methods for approximate search (Hidden) Markov models Why consider probabilistics

More information

Probabilistic Robotics. Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics (S. Thurn et al.

Probabilistic Robotics. Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics (S. Thurn et al. robabilistic Robotics Slides from Autonomous Robots Siegwart and Nourbaksh Chapter 5 robabilistic Robotics S. Thurn et al. Today Overview of probability Representing uncertainty ropagation of uncertainty

More information

Recursive Bayes Filtering

Recursive Bayes Filtering Recursive Bayes Filtering CS485 Autonomous Robotics Amarda Shehu Fall 2013 Notes modified from Wolfram Burgard, University of Freiburg Physical Agents are Inherently Uncertain Uncertainty arises from four

More information

HMM part 1. Dr Philip Jackson

HMM part 1. Dr Philip Jackson Centre for Vision Speech & Signal Processing University of Surrey, Guildford GU2 7XH. HMM part 1 Dr Philip Jackson Probability fundamentals Markov models State topology diagrams Hidden Markov models -

More information

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Wolfram Burgard, Cyrill Stachniss, Maren Bennewitz, Kai Arras 1 Motivation Recall: Discrete filter Discretize the

More information

PROBABILITY. Inference: constraint propagation

PROBABILITY. Inference: constraint propagation PROBABILITY Inference: constraint propagation! Use the constraints to reduce the number of legal values for a variable! Possible to find a solution without searching! Node consistency " A node is node-consistent

More information

CS37300 Class Notes. Jennifer Neville, Sebastian Moreno, Bruno Ribeiro

CS37300 Class Notes. Jennifer Neville, Sebastian Moreno, Bruno Ribeiro CS37300 Class Notes Jennifer Neville, Sebastian Moreno, Bruno Ribeiro 2 Background on Probability and Statistics These are basic definitions, concepts, and equations that should have been covered in your

More information

Autonomous Mobile Robot Design

Autonomous Mobile Robot Design Autonomous Mobile Robot Design Topic: Particle Filter for Localization Dr. Kostas Alexis (CSE) These slides relied on the lectures from C. Stachniss, and the book Probabilistic Robotics from Thurn et al.

More information

n(1 p i ) n 1 p i = 1 3 i=1 E(X i p = p i )P(p = p i ) = 1 3 p i = n 3 (p 1 + p 2 + p 3 ). p i i=1 P(X i = 1 p = p i )P(p = p i ) = p1+p2+p3

n(1 p i ) n 1 p i = 1 3 i=1 E(X i p = p i )P(p = p i ) = 1 3 p i = n 3 (p 1 + p 2 + p 3 ). p i i=1 P(X i = 1 p = p i )P(p = p i ) = p1+p2+p3 Introduction to Probability Due:August 8th, 211 Solutions of Final Exam Solve all the problems 1. (15 points) You have three coins, showing Head with probabilities p 1, p 2 and p 3. You perform two different

More information

Markov Models and Hidden Markov Models

Markov Models and Hidden Markov Models Markov Models and Hidden Markov Models Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. RN, AIMA Markov Models We have already seen that an MDP provides

More information

Why Probability? It's the right way to look at the world.

Why Probability? It's the right way to look at the world. Probability Why Probability? It's the right way to look at the world. Discrete Random Variables We denote discrete random variables with capital letters. A boolean random variable may be either true or

More information

Lecture 5: Random Walks and Markov Chain

Lecture 5: Random Walks and Markov Chain Spectral Graph Theory and Applications WS 20/202 Lecture 5: Random Walks and Markov Chain Lecturer: Thomas Sauerwald & He Sun Introduction to Markov Chains Definition 5.. A sequence of random variables

More information

We Live in Exciting Times. CSCI-567: Machine Learning (Spring 2019) Outline. Outline. ACM (an international computing research society) has named

We Live in Exciting Times. CSCI-567: Machine Learning (Spring 2019) Outline. Outline. ACM (an international computing research society) has named We Live in Exciting Times ACM (an international computing research society) has named CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Apr. 2, 2019 Yoshua Bengio,

More information

Approximate Inference

Approximate Inference Approximate Inference Simulation has a name: sampling Sampling is a hot topic in machine learning, and it s really simple Basic idea: Draw N samples from a sampling distribution S Compute an approximate

More information

18.175: Lecture 30 Markov chains

18.175: Lecture 30 Markov chains 18.175: Lecture 30 Markov chains Scott Sheffield MIT Outline Review what you know about finite state Markov chains Finite state ergodicity and stationarity More general setup Outline Review what you know

More information

Dept. of Linguistics, Indiana University Fall 2015

Dept. of Linguistics, Indiana University Fall 2015 L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 34 To start out the course, we need to know something about statistics and This is only an introduction; for a fuller understanding, you would

More information

Hidden Markov Models. AIMA Chapter 15, Sections 1 5. AIMA Chapter 15, Sections 1 5 1

Hidden Markov Models. AIMA Chapter 15, Sections 1 5. AIMA Chapter 15, Sections 1 5 1 Hidden Markov Models AIMA Chapter 15, Sections 1 5 AIMA Chapter 15, Sections 1 5 1 Consider a target tracking problem Time and uncertainty X t = set of unobservable state variables at time t e.g., Position

More information

Learning Sequence Motif Models Using Expectation Maximization (EM) and Gibbs Sampling

Learning Sequence Motif Models Using Expectation Maximization (EM) and Gibbs Sampling Learning Sequence Motif Models Using Expectation Maximization (EM) and Gibbs Sampling BMI/CS 776 www.biostat.wisc.edu/bmi776/ Spring 009 Mark Craven craven@biostat.wisc.edu Sequence Motifs what is a sequence

More information

Notes on the second moment method, Erdős multiplication tables

Notes on the second moment method, Erdős multiplication tables Notes on the second moment method, Erdős multiplication tables January 25, 20 Erdős multiplication table theorem Suppose we form the N N multiplication table, containing all the N 2 products ab, where

More information

Results: MCMC Dancers, q=10, n=500

Results: MCMC Dancers, q=10, n=500 Motivation Sampling Methods for Bayesian Inference How to track many INTERACTING targets? A Tutorial Frank Dellaert Results: MCMC Dancers, q=10, n=500 1 Probabilistic Topological Maps Results Real-Time

More information

Sampling Based Filters

Sampling Based Filters Statistical Techniques in Robotics (16-831, F12) Lecture#03 (Monday September 5) Sampling Based Filters Lecturer: Drew Bagnell Scribes: M. Taylor, M. Shomin 1 1 Beam-based Sensor Model Many common sensors

More information

Entropy Rate of Stochastic Processes

Entropy Rate of Stochastic Processes Entropy Rate of Stochastic Processes Timo Mulder tmamulder@gmail.com Jorn Peters jornpeters@gmail.com February 8, 205 The entropy rate of independent and identically distributed events can on average be

More information

Math Camp Notes: Linear Algebra II

Math Camp Notes: Linear Algebra II Math Camp Notes: Linear Algebra II Eigenvalues Let A be a square matrix. An eigenvalue is a number λ which when subtracted from the diagonal elements of the matrix A creates a singular matrix. In other

More information

Hidden Markov Models. Vibhav Gogate The University of Texas at Dallas

Hidden Markov Models. Vibhav Gogate The University of Texas at Dallas Hidden Markov Models Vibhav Gogate The University of Texas at Dallas Intro to AI (CS 4365) Many slides over the course adapted from either Dan Klein, Luke Zettlemoyer, Stuart Russell or Andrew Moore 1

More information

Gaussian Processes for Sequential Prediction

Gaussian Processes for Sequential Prediction Gaussian Processes for Sequential Prediction Michael A. Osborne Machine Learning Research Group Department of Engineering Science University of Oxford Gaussian processes are useful for sequential data,

More information

CSE 473: Artificial Intelligence

CSE 473: Artificial Intelligence CSE 473: Artificial Intelligence Hidden Markov Models Dieter Fox --- University of Washington [Most slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials

More information

Sampling Based Filters

Sampling Based Filters Statistical Techniques in Robotics (16-831, F11) Lecture#03 (Wednesday September 7) Sampling Based Filters Lecturer: Drew Bagnell Scribe:Nate Wood 1 1 Monte Carlo Method As discussed in previous lectures,

More information

IE 336 Seat # Name. Closed book. One page of hand-written notes, front and back. No calculator. 60 minutes.

IE 336 Seat # Name. Closed book. One page of hand-written notes, front and back. No calculator. 60 minutes. Closed book. One page of hand-written notes, front and back. No calculator. 60 minutes. Cover page and five pages of exam. Four questions. To receive full credit, show enough work to indicate your logic.

More information

Infering the Number of State Clusters in Hidden Markov Model and its Extension

Infering the Number of State Clusters in Hidden Markov Model and its Extension Infering the Number of State Clusters in Hidden Markov Model and its Extension Xugang Ye Department of Applied Mathematics and Statistics, Johns Hopkins University Elements of a Hidden Markov Model (HMM)

More information

Hidden Markov models 1

Hidden Markov models 1 Hidden Markov models 1 Outline Time and uncertainty Markov process Hidden Markov models Inference: filtering, prediction, smoothing Most likely explanation: Viterbi 2 Time and uncertainty The world changes;

More information

Our Status in CSE 5522

Our Status in CSE 5522 Our Status in CSE 5522 We re done with Part I Search and Planning! Part II: Probabilistic Reasoning Diagnosis Speech recognition Tracking objects Robot mapping Genetics Error correcting codes lots more!

More information

Today. Next lecture. (Ch 14) Markov chains and hidden Markov models

Today. Next lecture. (Ch 14) Markov chains and hidden Markov models Today (Ch 14) Markov chains and hidden Markov models Graphical representation Transition probability matrix Propagating state distributions The stationary distribution Next lecture (Ch 14) Markov chains

More information

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes.

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes. CMSC 310 Artificial Intelligence Probabilistic Reasoning and Bayesian Belief Networks Probabilities, Random Variables, Probability Distribution, Conditional Probability, Joint Distributions, Bayes Theorem

More information

CS188 Outline. We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning. Part III: Machine Learning

CS188 Outline. We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning. Part III: Machine Learning CS188 Outline We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning Diagnosis Speech recognition Tracking objects Robot mapping Genetics Error correcting codes lots more! Part III:

More information

Sheet 4 solutions. May 15, 2017

Sheet 4 solutions. May 15, 2017 Sheet 4 solutions May 15, 2017 Exercise 1: Bayes Rule Suppose you are a witness to a nighttime hit-and-run accident involving a taxi in Athens. All taxi cars in Athens are blue or green. You swear under

More information

Particle Filters. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics

Particle Filters. Pieter Abbeel UC Berkeley EECS. Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Particle Filters Pieter Abbeel UC Berkeley EECS Many slides adapted from Thrun, Burgard and Fox, Probabilistic Robotics Motivation For continuous spaces: often no analytical formulas for Bayes filter updates

More information

The Naïve Bayes Classifier. Machine Learning Fall 2017

The Naïve Bayes Classifier. Machine Learning Fall 2017 The Naïve Bayes Classifier Machine Learning Fall 2017 1 Today s lecture The naïve Bayes Classifier Learning the naïve Bayes Classifier Practical concerns 2 Today s lecture The naïve Bayes Classifier Learning

More information

Earth Science: Second Quarter Grading Rubric Kindergarten

Earth Science: Second Quarter Grading Rubric Kindergarten Earth Science: Second Quarter Grading Rubric Kindergarten of their senses Observation skills are used to note characteristics of our environment on a daily basis. The weather may be sunny one day and cloudy

More information

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 18, 2015 1 / 45 Resources and Attribution Image credits,

More information

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric

More information

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST Information Theory Lecture 5 Entropy rate and Markov sources STEFAN HÖST Universal Source Coding Huffman coding is optimal, what is the problem? In the previous coding schemes (Huffman and Shannon-Fano)it

More information

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 8: Importance Sampling

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 8: Importance Sampling Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 8: Importance Sampling 8.1 Importance Sampling Importance sampling

More information

COMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma

COMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma COMS 4771 Probabilistic Reasoning via Graphical Models Nakul Verma Last time Dimensionality Reduction Linear vs non-linear Dimensionality Reduction Principal Component Analysis (PCA) Non-linear methods

More information

Machine Learning. CS Spring 2015 a Bayesian Learning (I) Uncertainty

Machine Learning. CS Spring 2015 a Bayesian Learning (I) Uncertainty Machine Learning CS6375 --- Spring 2015 a Bayesian Learning (I) 1 Uncertainty Most real-world problems deal with uncertain information Diagnosis: Likely disease given observed symptoms Equipment repair:

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Oct, 21, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models CPSC

More information

Grundlagen der Künstlichen Intelligenz

Grundlagen der Künstlichen Intelligenz Grundlagen der Künstlichen Intelligenz Uncertainty & Probabilities & Bandits Daniel Hennes 16.11.2017 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Uncertainty Probability

More information

Informatics 2D Reasoning and Agents Semester 2,

Informatics 2D Reasoning and Agents Semester 2, Informatics 2D Reasoning and Agents Semester 2, 2018 2019 Alex Lascarides alex@inf.ed.ac.uk Lecture 25 Approximate Inference in Bayesian Networks 19th March 2019 Informatics UoE Informatics 2D 1 Where

More information

Reasoning Under Uncertainty: Conditional Probability

Reasoning Under Uncertainty: Conditional Probability Reasoning Under Uncertainty: Conditional Probability CPSC 322 Uncertainty 2 Textbook 6.1 Reasoning Under Uncertainty: Conditional Probability CPSC 322 Uncertainty 2, Slide 1 Lecture Overview 1 Recap 2

More information

UCSD ECE250 Handout #20 Prof. Young-Han Kim Monday, February 26, Solutions to Exercise Set #7

UCSD ECE250 Handout #20 Prof. Young-Han Kim Monday, February 26, Solutions to Exercise Set #7 UCSD ECE50 Handout #0 Prof. Young-Han Kim Monday, February 6, 07 Solutions to Exercise Set #7. Minimum waiting time. Let X,X,... be i.i.d. exponentially distributed random variables with parameter λ, i.e.,

More information

CS Homework 2: Combinatorics & Discrete Events Due Date: September 25, 2018 at 2:20 PM

CS Homework 2: Combinatorics & Discrete Events Due Date: September 25, 2018 at 2:20 PM CS1450 - Homework 2: Combinatorics & Discrete Events Due Date: September 25, 2018 at 2:20 PM Question 1 A website allows the user to create an 8-character password that consists of lower case letters (a-z)

More information

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech Probabilistic Graphical Models and Bayesian Networks Artificial Intelligence Bert Huang Virginia Tech Concept Map for Segment Probabilistic Graphical Models Probabilistic Time Series Models Particle Filters

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression

More information

Linear Algebra Solutions 1

Linear Algebra Solutions 1 Math Camp 1 Do the following: Linear Algebra Solutions 1 1. Let A = and B = 3 8 5 A B = 3 5 9 A + B = 9 11 14 4 AB = 69 3 16 BA = 1 4 ( 1 3. Let v = and u = 5 uv = 13 u v = 13 v u = 13 Math Camp 1 ( 7

More information

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:

More information

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes CDA6530: Performance Models of Computers and Networks Chapter 3: Review of Practical Stochastic Processes Definition Stochastic process X = {X(t), t2 T} is a collection of random variables (rvs); one rv

More information

CDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical

CDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical CDA5530: Performance Models of Computers and Networks Chapter 3: Review of Practical Stochastic Processes Definition Stochastic ti process X = {X(t), t T} is a collection of random variables (rvs); one

More information

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)

More information

Machine Learning. Yuh-Jye Lee. March 1, Lab of Data Science and Machine Intelligence Dept. of Applied Math. at NCTU

Machine Learning. Yuh-Jye Lee. March 1, Lab of Data Science and Machine Intelligence Dept. of Applied Math. at NCTU Machine Learning Yuh-Jye Lee Lab of Data Science and Machine Intelligence Dept. of Applied Math. at NCTU March 1, 2017 1 / 13 Bayes Rule Bayes Rule Assume that {B 1, B 2,..., B k } is a partition of S

More information

There are two basic kinds of random variables continuous and discrete.

There are two basic kinds of random variables continuous and discrete. Summary of Lectures 5 and 6 Random Variables The random variable is usually represented by an upper case letter, say X. A measured value of the random variable is denoted by the corresponding lower case

More information

Discrete Random Variables

Discrete Random Variables Probability Discrete Random Variables We denote discrete random variables with capital letters. A boolean random variable may be either true or false A = true or A=false. P(a), or P(A=true) denotes the

More information

Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule

Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule Alan Mackworth UBC CS 322 Uncertainty 2 March 13, 2013 Textbook 6.1.3 Lecture Overview Recap: Probability & Possible World Semantics

More information

THE QUEEN S UNIVERSITY OF BELFAST

THE QUEEN S UNIVERSITY OF BELFAST THE QUEEN S UNIVERSITY OF BELFAST 0SOR20 Level 2 Examination Statistics and Operational Research 20 Probability and Distribution Theory Wednesday 4 August 2002 2.30 pm 5.30 pm Examiners { Professor R M

More information

CS 630 Basic Probability and Information Theory. Tim Campbell

CS 630 Basic Probability and Information Theory. Tim Campbell CS 630 Basic Probability and Information Theory Tim Campbell 21 January 2003 Probability Theory Probability Theory is the study of how best to predict outcomes of events. An experiment (or trial or event)

More information

O June, 2010 MMT-008 : PROBABILITY AND STATISTICS

O June, 2010 MMT-008 : PROBABILITY AND STATISTICS No. of Printed Pages : 8 M.Sc. MATHEMATICS WITH APPLICATIONS IN COMPUTER SCIENCE (MACS) tr.) Term-End Examination O June, 2010 : PROBABILITY AND STATISTICS Time : 3 hours Maximum Marks : 100 Note : Question

More information

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that

More information

Linear Dynamical Systems (Kalman filter)

Linear Dynamical Systems (Kalman filter) Linear Dynamical Systems (Kalman filter) (a) Overview of HMMs (b) From HMMs to Linear Dynamical Systems (LDS) 1 Markov Chains with Discrete Random Variables x 1 x 2 x 3 x T Let s assume we have discrete

More information

CS 188: Artificial Intelligence Fall 2011

CS 188: Artificial Intelligence Fall 2011 CS 188: Artificial Intelligence Fall 2011 Lecture 12: Probability 10/4/2011 Dan Klein UC Berkeley 1 Today Probability Random Variables Joint and Marginal Distributions Conditional Distribution Product

More information

CS 343: Artificial Intelligence

CS 343: Artificial Intelligence CS 343: Artificial Intelligence Probability Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188

More information

Fusion in simple models

Fusion in simple models Fusion in simple models Peter Antal antal@mit.bme.hu A.I. February 8, 2018 1 Basic concepts of probability theory Joint distribution Conditional probability Bayes rule Chain rule Marginalization General

More information

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303)

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

CS 188: Artificial Intelligence. Bayes Nets

CS 188: Artificial Intelligence. Bayes Nets CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew

More information

Introduction to Probability and Statistics (Continued)

Introduction to Probability and Statistics (Continued) Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:

More information

COMPSCI 240: Reasoning Under Uncertainty

COMPSCI 240: Reasoning Under Uncertainty COMPSCI 240: Reasoning Under Uncertainty Andrew Lan and Nic Herndon University of Massachusetts at Amherst Spring 2019 Lecture 20: Central limit theorem & The strong law of large numbers Markov and Chebyshev

More information

01 Probability Theory and Statistics Review

01 Probability Theory and Statistics Review NAVARCH/EECS 568, ROB 530 - Winter 2018 01 Probability Theory and Statistics Review Maani Ghaffari January 08, 2018 Last Time: Bayes Filters Given: Stream of observations z 1:t and action data u 1:t Sensor/measurement

More information

CSEP 573: Artificial Intelligence

CSEP 573: Artificial Intelligence CSEP 573: Artificial Intelligence Hidden Markov Models Luke Zettlemoyer Many slides over the course adapted from either Dan Klein, Stuart Russell, Andrew Moore, Ali Farhadi, or Dan Weld 1 Outline Probabilistic

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning CS4375 --- Fall 2018 Bayesian a Learning Reading: Sections 13.1-13.6, 20.1-20.2, R&N Sections 6.1-6.3, 6.7, 6.9, Mitchell 1 Uncertainty Most real-world problems deal with

More information

Sequential Data and Markov Models

Sequential Data and Markov Models equential Data and Markov Models argur N. rihari srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/ce574/index.html 0 equential Data Examples Often arise through

More information

What is a random variable

What is a random variable OKAN UNIVERSITY FACULTY OF ENGINEERING AND ARCHITECTURE MATH 256 Probability and Random Processes 04 Random Variables Fall 20 Yrd. Doç. Dr. Didem Kivanc Tureli didemk@ieee.org didem.kivanc@okan.edu.tr

More information

Simultaneous Localization and Mapping

Simultaneous Localization and Mapping Simultaneous Localization and Mapping Miroslav Kulich Intelligent and Mobile Robotics Group Czech Institute of Informatics, Robotics and Cybernetics Czech Technical University in Prague Winter semester

More information

CS 188: Artificial Intelligence Spring 2009

CS 188: Artificial Intelligence Spring 2009 CS 188: Artificial Intelligence Spring 2009 Lecture 21: Hidden Markov Models 4/7/2009 John DeNero UC Berkeley Slides adapted from Dan Klein Announcements Written 3 deadline extended! Posted last Friday

More information

MATH3200, Lecture 31: Applications of Eigenvectors. Markov Chains and Chemical Reaction Systems

MATH3200, Lecture 31: Applications of Eigenvectors. Markov Chains and Chemical Reaction Systems Lecture 31: Some Applications of Eigenvectors: Markov Chains and Chemical Reaction Systems Winfried Just Department of Mathematics, Ohio University April 9 11, 2018 Review: Eigenvectors and left eigenvectors

More information

Introduction to Machine Learning

Introduction to Machine Learning Uncertainty Introduction to Machine Learning CS4375 --- Fall 2018 a Bayesian Learning Reading: Sections 13.1-13.6, 20.1-20.2, R&N Sections 6.1-6.3, 6.7, 6.9, Mitchell Most real-world problems deal with

More information

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods Prof. Daniel Cremers 14. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric

More information

Markov Processes Hamid R. Rabiee

Markov Processes Hamid R. Rabiee Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete

More information

Probability Theory. Introduction to Probability Theory. Principles of Counting Examples. Principles of Counting. Probability spaces.

Probability Theory. Introduction to Probability Theory. Principles of Counting Examples. Principles of Counting. Probability spaces. Probability Theory To start out the course, we need to know something about statistics and probability Introduction to Probability Theory L645 Advanced NLP Autumn 2009 This is only an introduction; for

More information

Quantitative Information Flow. Lecture 7

Quantitative Information Flow. Lecture 7 Quantitative Information Flow Lecture 7 1 The basic model: Systems = Information-Theoretic channels Secret Information Observables s1 o1... System... sm on Input Output 2 Probabilistic systems are noisy

More information

Supervised Learning Hidden Markov Models. Some of these slides were inspired by the tutorials of Andrew Moore

Supervised Learning Hidden Markov Models. Some of these slides were inspired by the tutorials of Andrew Moore Supervised Learning Hidden Markov Models Some of these slides were inspired by the tutorials of Andrew Moore A Markov System S 2 Has N states, called s 1, s 2.. s N There are discrete timesteps, t=0, t=1,.

More information

SLAM Techniques and Algorithms. Jack Collier. Canada. Recherche et développement pour la défense Canada. Defence Research and Development Canada

SLAM Techniques and Algorithms. Jack Collier. Canada. Recherche et développement pour la défense Canada. Defence Research and Development Canada SLAM Techniques and Algorithms Jack Collier Defence Research and Development Canada Recherche et développement pour la défense Canada Canada Goals What will we learn Gain an appreciation for what SLAM

More information