CS 188: Artificial Intelligence Fall Announcements

Similar documents
Today. CS 188: Artificial Intelligence. Recap: Reasoning Over Time. Particle Filters and Applications of HMMs. HMMs

Reasoning over Time or Space. CS 188: Artificial Intelligence. Outline. Markov Models. Conditional Independence. Query: P(X 4 )

Today. Recap: Reasoning Over Time. Demo Bonanza! CS 188: Artificial Intelligence. Advanced HMMs. Speech recognition. HMMs. Start machine learning

Pacman in the News. Announcements. Pacman in the News. CS 188: Ar7ficial Intelligence Par7cle Filters and A pplica7ons of HMMs

CS 343: Artificial Intelligence

CS 5522: Artificial Intelligence II

CS 5522: Artificial Intelligence II

Decision Networks. CS 188: Artificial Intelligence Fall Example: Decision Networks. Decision Networks. Decisions as Outcome Trees

CS 343: Artificial Intelligence

Probabilistic Reasoning. CS 188: Artificial Intelligence Spring Inference by Enumeration. Probability recap. Chain Rule à Bayes net

Discrete Mathematics and Probability Theory Summer 2014 James Cook Note 17

CS 188: Artificial Intelligence Spring Announcements

Bayesian Networks: Approximate Inference

Markov Models. CS 188: Artificial Intelligence Fall Example. Mini-Forward Algorithm. Stationary Distributions.

Decision Networks. CS 188: Artificial Intelligence. Decision Networks. Decision Networks. Decision Networks and Value of Information

Par+cle Filtering. CSE 473: Ar+ficial Intelligence Par+cle Filters. Par+cle Filtering: Elapse Time. Representa+on: Par+cles

CS 188: Artificial Intelligence

CS 188: Artificial Intelligence Fall 2010

CSEP 573: Artificial Intelligence

Discrete Mathematics and Probability Theory Spring 2013 Anant Sahai Lecture 17

Reinforcement learning II

Announcements. CS 188: Artificial Intelligence Fall VPI Example. VPI Properties. Reasoning over Time. Markov Models. Lecture 19: HMMs 11/4/2008

Review of Gaussian Quadrature method

CS103B Handout 18 Winter 2007 February 28, 2007 Finite Automata

CS 188: Artificial Intelligence

Convert the NFA into DFA

Z b. f(x)dx. Yet in the above two cases we know what f(x) is. Sometimes, engineers want to calculate an area by computing I, but...

CSE 473: Ar+ficial Intelligence. Example. Par+cle Filters for HMMs. An HMM is defined by: Ini+al distribu+on: Transi+ons: Emissions:

Markov Models and Hidden Markov Models

Module 6 Value Iteration. CS 886 Sequential Decision Making and Reinforcement Learning University of Waterloo

Math 520 Final Exam Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

CSE 473: Ar+ficial Intelligence. Probability Recap. Markov Models - II. Condi+onal probability. Product rule. Chain rule.

CS 275 Automata and Formal Language Theory

CS 373, Spring Solutions to Mock midterm 1 (Based on first midterm in CS 273, Fall 2008.)

Announcements. CS 188: Artificial Intelligence Fall Markov Models. Example: Markov Chain. Mini-Forward Algorithm. Example

Minimal DFA. minimal DFA for L starting from any other

Hidden Markov Models. Vibhav Gogate The University of Texas at Dallas

Probability, Markov models and HMMs. Vibhav Gogate The University of Texas at Dallas

Hidden Markov Models. Hal Daumé III. Computer Science University of Maryland CS 421: Introduction to Artificial Intelligence 19 Apr 2012

Some Theory of Computation Exercises Week 1

CS667 Lecture 6: Monte Carlo Integration 02/10/05

CSE 473: Ar+ficial Intelligence

Quantum Nonlocality Pt. 2: No-Signaling and Local Hidden Variables May 1, / 16

Lecture 3. In this lecture, we will discuss algorithms for solving systems of linear equations.

Chapter 3. Vector Spaces

1B40 Practical Skills

Nondeterminism and Nodeterministic Automata

Genetic Programming. Outline. Evolutionary Strategies. Evolutionary strategies Genetic programming Summary

Non-Deterministic Finite Automata. Fall 2018 Costas Busch - RPI 1

1 From NFA to regular expression

LECTURE NOTE #12 PROF. ALAN YUILLE

State space systems analysis (continued) Stability. A. Definitions A system is said to be Asymptotically Stable (AS) when it satisfies

Bases for Vector Spaces

CS 301. Lecture 04 Regular Expressions. Stephen Checkoway. January 29, 2018

APPROXIMATE INTEGRATION

Flexible Beam. Objectives

CSCI 340: Computational Models. Kleene s Theorem. Department of Computer Science

How do we solve these things, especially when they get complicated? How do we know when a system has a solution, and when is it unique?

Lecture 08: Feb. 08, 2019

{ } = E! & $ " k r t +k +1

Lecture 6. Notes. Notes. Notes. Representations Z A B and A B R. BTE Electronics Fundamentals August Bern University of Applied Sciences

Lecture 9: LTL and Büchi Automata

Chapter 4: Dynamic Programming

AQA Further Pure 1. Complex Numbers. Section 1: Introduction to Complex Numbers. The number system

QUADRATURE is an old-fashioned word that refers to

Session 13

5. (±±) Λ = fw j w is string of even lengthg [ 00 = f11,00g 7. (11 [ 00)± Λ = fw j w egins with either 11 or 00g 8. (0 [ ffl)1 Λ = 01 Λ [ 1 Λ 9.

Lexical Analysis Finite Automate

Non-deterministic Finite Automata

NFAs continued, Closure Properties of Regular Languages

1.4 Nonregular Languages

Intermediate Math Circles Wednesday, November 14, 2018 Finite Automata II. Nickolas Rollick a b b. a b 4

Chapter 5 Plan-Space Planning

Particle Filtering. CSE 473: Artificial Intelligence Particle Filters. Representation: Particles. Particle Filtering: Elapse Time

19 Optimal behavior: Game theory

CMPSCI 250: Introduction to Computation. Lecture #31: What DFA s Can and Can t Do David Mix Barrington 9 April 2014

CS 188: Artificial Intelligence Spring 2007

Construction of Gauss Quadrature Rules

Chapter 14. Matrix Representations of Linear Transformations

Lecture Solution of a System of Linear Equation

Review of Calculus, cont d

Section: Other Models of Turing Machines. Definition: Two automata are equivalent if they accept the same language.

Continuous Random Variables

Jim Lambers MAT 169 Fall Semester Lecture 4 Notes

Math 426: Probability Final Exam Practice

Lecture 3: Curves in Calculus. Table of contents

Monte Carlo method in solving numerical integration and differential equation

Administrivia CSE 190: Reinforcement Learning: An Introduction

Section 6.1 Definite Integral

The final exam will take place on Friday May 11th from 8am 11am in Evans room 60.

Chapters Five Notes SN AA U1C5

Suppose we want to find the area under the parabola and above the x axis, between the lines x = 2 and x = -2.

C/CS/Phys C191 Bell Inequalities, No Cloning, Teleportation 9/13/07 Fall 2007 Lecture 6

Lecture 2e Orthogonal Complement (pages )

NFA DFA Example 3 CMSC 330: Organization of Programming Languages. Equivalence of DFAs and NFAs. Equivalence of DFAs and NFAs (cont.

Driving Cycle Construction of City Road for Hybrid Bus Based on Markov Process Deng Pan1, a, Fengchun Sun1,b*, Hongwen He1, c, Jiankun Peng1, d

MTH 122 Fall 2008 Essex County College Division of Mathematics Handout Version 10 1 October 14, 2008

11.1 Finite Automata. CS125 Lecture 11 Fall Motivation: TMs without a tape: maybe we can at least fully understand such a simple model?

Name Ima Sample ASU ID

Harvard University Computer Science 121 Midterm October 23, 2012

Transcription:

CS 188: Artificil Intelligence Fll 2009 Lecture 20: Prticle Filtering 11/5/2009 Dn Klein UC Berkeley Announcements Written 3 out: due 10/12 Project 4 out: due 10/19 Written 4 proly xed, Project 5 moving up Course contest updte Dily tournments re now eing run! Instructions for registrtion on the wesite Qulifying opens soon: 1% on finl exm! 2 1

Recp: Resoning Over Time Sttionry Mrkov models 0.3 X 1 X 2 X 3 X 4 rin sun 0.7 0.3 0.7 Hidden Mrkov models X 1 X 2 X 3 X 4 X 5 E 2 E 3 E 4 E 5 X E P rin umrell 0.9 rin no umrell 0.1 sun umrell 0.2 sun no umrell 0.8 Recp: Filtering Elpse time: compute P( X t e 1:t-1 ) Oserve: compute P( X t e 1:t ) X 1 X 2 Belief: <P(rin), P(sun)> <0.5, 0.5> Prior on X 1 <0.82, 0.18> Oserve E 2 <0.63, 0.37> Elpse time <0.88, 0.12> Oserve 2

Prticle Filtering Filtering: pproximte solution Sometimes X is too ig to use exct inference X my e too ig to even store B(X) E.g. X is continuous Solution: pproximte inference Trck smples of X, not ll vlues Smples re clled prticles Time per step is liner in the numer of smples But: numer needed my e lrge In memory: list of prticles, not sttes This is how root locliztion works in prctice 0.0 0.1 0.0 0.0 0.0 0.2 0.0 0.2 0.5 Prticle Filtering: Elpse Time Ech prticle is moved y smpling its next position from the trnsition model This is like prior smpling smples frequencies reflect the trnsition pros Here, most smples move clockwise, ut some move in nother direction or sty in plce This cptures the pssge of time If we hve enough smples, close to the exct vlues efore nd fter (consistent) 3

Prticle Filtering: Oserve Slightly trickier: We don t smple the oservtion, we fix it This is similr to likelihood weighting, so we downweight our smples sed on the evidence Note tht, s efore, the proilities don t sum to one, since most hve een downweighted (in fct they sum to n pproximtion of P(e)) Prticle Filtering: Resmple Rther thn trcking weighted smples, we resmple N times, we choose from our weighted smple distriution (i.e. drw with replcement) This is equivlent to renormlizing the distriution Now the updte is complete for this time step, continue with the next one 4

Prticle Filtering Sometimes X is too ig to use exct inference X my e too ig to even store B(X) E.g. X is continuous X 2 my e too ig to do updtes Solution: pproximte inference Trck smples of X, not ll vlues Smples re clled prticles Time per step is liner in the numer of smples But: numer needed my e lrge In memory: list of prticles, not sttes 0.0 0.1 0.0 0.0 0.0 0.2 0.0 0.2 0.5 This is how root locliztion works in prctice Representtion: Prticles Our representtion of P(X) is now list of N prticles (smples) Generlly, N << X Storing mp from X to counts would defet the point P(x) pproximted y numer of prticles with vlue x So, mny x will hve P(x) = 0! More prticles, more ccurcy For now, ll prticles hve weight of 1 Prticles: (3,3) (2,3) (3,3) (3,2) (3,3) (3,2) (2,1) (3,3) (3,3) (2,1) 15 5

Prticle Filtering: Elpse Time Ech prticle is moved y smpling its next position from the trnsition model This is like prior smpling smples frequencies reflect the trnsition pros Here, most smples move clockwise, ut some move in nother direction or sty in plce This cptures the pssge of time If we hve enough smples, close to the exct vlues efore nd fter (consistent) Prticle Filtering: Oserve Slightly trickier: Don t do rejection smpling (why not?) We don t smple the oservtion, we fix it This is similr to likelihood weighting, so we downweight our smples sed on the evidence Note tht, s efore, the proilities don t sum to one, since most hve een downweighted (in fct they sum to n pproximtion of P(e)) 6

Prticle Filtering: Resmple Rther thn trcking weighted smples, we resmple N times, we choose from our weighted smple distriution (i.e. drw with replcement) This is equivlent to renormlizing the distriution Now the updte is complete for this time step, continue with the next one Old Prticles: (3,3) w=0.1 (2,1) w=0.9 (2,1) w=0.9 (3,1) w=0.4 (3,2) w=0.3 (2,2) w=0.4 (1,1) w=0.4 (3,1) w=0.4 (2,1) w=0.9 (3,2) w=0.3 Old Prticles: (2,1) w=1 (2,1) w=1 (2,1) w=1 (3,2) w=1 (2,2) w=1 (2,1) w=1 (1,1) w=1 (3,1) w=1 (2,1) w=1 (1,1) w=1 In root locliztion: Root Locliztion We know the mp, ut not the root s position Oservtions my e vectors of rnge finder redings Stte spce nd redings re typiclly continuous (works siclly like very fine grid) nd so we cnnot store B(X) Prticle filtering is min technique [Demos] 7

P4: Ghostusters 2.0 (et) Plot: Pcmn's grndfther, Grndpc, lerned to hunt ghosts for sport. He ws linded y his power, ut could her the ghosts nging nd clnging. Trnsition Model: All ghosts move rndomly, ut re sometimes ised Noisy distnce pro True distnce = 8 15 13 11 9 7 5 Emission Model: Pcmn knows noisy distnce to ech ghost 3 1 [Demo] Dynmic Byes Nets (DBNs) We wnt to trck multiple vriles over time, using multiple sources of evidence Ide: Repet fixed Byes net structure t ech time Vriles from time t cn condition on those from t-1 t =1 t =2 t =3 G 1 G 2 G 3 G 1 G 2 G 3 E 2 E 2 E 3 E 3 Discrete vlued dynmic Byes nets re lso HMMs 8

Exct Inference in DBNs Vrile elimintion pplies to dynmic Byes nets Procedure: unroll the network for T time steps, then eliminte vriles until P(X T e 1:T ) is computed t =1 t =2 t =3 G 1 G 2 G 3 G 1 G 2 G 3 E 2 E 2 E 3 E 3 Online elief updtes: Eliminte ll vriles from the previous time step; store fctors for current time only 22 DBN Prticle Filters A prticle is complete smple for time step Initilize: Generte prior smples for the t=1 Byes net Exmple prticle: G 1 = (3,3) G 1 = (5,3) Elpse time: Smple successor for ech prticle Exmple successor: G 2 = (2,3) G 2 = (6,3) Oserve: Weight ech entire smple y the likelihood of the evidence conditioned on the smple Likelihood: P( G 1 ) * P( G 1 ) Resmple: Select prior smples (tuples of vlues) in proportion to their likelihood [Demo] 23 9

SLAM SLAM = Simultneous Locliztion And Mpping We do not know the mp or our loction Our elief stte is over mps nd positions! Min techniques: Klmn filtering (Gussin HMMs) nd prticle methods [DEMOS] DP-SLAM, Ron Prr 10