Computing Posterior Probabilities. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington
|
|
- Letitia Goodwin
- 5 years ago
- Views:
Transcription
1 Computing Posterior Probabilities Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington 1
2 Overview of Candy Bag Example As described in Russell and Norvig, for Chapter 20 of the 2 nd edition: Five kinds of bags of candies. 10% are h 1 : 100% cherry candies 20% are h 2 : 75% cherry candies + 25% lime candies 40% are h 3 : 50% cherry candies + 50% lime candies 20% are h 4 : 25% cherry candies + 75% lime candies 10% are h 5 : 100% lime candies Each bag has an infinite number of candies. This way, the ratio of candy types inside a bag does not change as we pick candies out of the bag. We have a bag, and we are picking candies out of it. Based on the types of candies we are picking, we want to figure out what type of bag we have. 2
3 Hypotheses and Prior Probabilities Five kinds of bags of candies. 10% are h 1 : 100% cherry candies 20% are h 2 : 75% cherry candies + 25% lime candies 40% are h 3 : 50% cherry candies + 50% lime candies 20% are h 4 : 25% cherry candies + 75% lime candies 10% are h 5 : 100% lime candies Each h i is called a hypothesis. The initial probability that is given for each hypothesis is called the prior probability for that hypothesis. It is called prior because it is the probability we have before we have made any observations. 3
4 Observations and Posteriors Out of our bag, we pick T candies, whose types are: Q 1, Q 2,, Q T. Each Q j is equal to either C (cherry) or L ( lime ). These Q j s are called the observations. Based on our observations, we want to answer two types of questions: What is P(h i Q 1,, Q t )? Probability of hypothesis i after t observations. This is called the posterior probability of h i. What is P(Q t+1 = C Q 1,, Q t )? Similarly, what is P(Q t+1 = L Q 1,, Q t ) Probability of observation t+1 after t observations. 4
5 Simplifying notation Define: P t (h i ) = P(h i Q 1,, Q t ) P t (Q t+1 = C) = P(Q t+1 = C Q 1,, Q t )? Special case: t = 0 (no observations): P 0 (h i ) = P(h i ) P 0 (h i ) is the prior probability of h i P 0 (Q 1 = C) = P(Q 1 = C) P 0 (Q 1 = C) is the probability that the first observation is equal to C. 5
6 Questions We Want to Answer, Revisited Using the simplified notation of the previous slide: What is P t (h i )? Posterior probability of hypothesis i after t observations. What is P t (Q t+1 = C)? Similarly, what is P t (Q t+1 = L) Probability of observation t+1 after t observations. 6
7 Computing P 0 (Q t ) As an example: Consider P 0 (Q 1 = C). What does P 0 (Q 1 = C) mean? 7
8 Computing P 0 (Q t ) As an example: Consider P 0 (Q 1 = C). What does P 0 (Q 1 = C) mean? It is the probability that the first candy we pick out of our bag is a cherry candy. P 0 Q 1 = C = P Q 1 = C h 1 ) P(h 1 ) + P Q 1 = C h 2 ) P(h 2 ) + P Q 1 = C h 3 ) P(h 3 ) + P Q 1 = C h 4 ) P(h 4 ) + P Q 1 = C h 5 ) P(h 5 ) 8
9 Computing P 0 (Q t ) As an example: Consider P 0 (Q 1 = C). What does P 0 (Q 1 = C) mean? It is the probability that the first candy we pick out of our bag is a cherry candy. P 0 Q 1 = C = = 0.5 9
10 Computing P 1 (h i ) As an example: Consider P 1 (h 1 ). What does P 1 (h 1 )mean? P 1 h 1 = P h 1 Q 1 = C) P 1 h 1 is the probability that our bag is of type h 1, if the first candy we pick out of our bag is a cherry candy. 10
11 Computing P 1 (h i ) As an example: Consider P 1 (h 1 ). P 1 h 1 = P h 1 Q 1 = C) P h 1 Q 1 = C) = 11
12 Computing P 1 (h i ) As an example: Consider P 1 (h 1 ). P 1 h 1 = P h 1 Q 1 = C) P h 1 Q 1 = C) = P Q 1=C h 1 ) P(h 1 ) P(Q 1 =C) 12
13 Computing P 1 (h i ) As an example: Consider P 1 (h 1 ). P 1 h 1 = P h 1 Q 1 = C) P h 1 Q 1 = C) = P Q 1=C h 1 ) P(h 1 ) P(Q 1 =C) = =
14 Computing P 1 (h i ) Consider P 1 (h 2 ). P 1 h 2 = P h 2 Q 1 = C) P h 2 Q 1 = C) = P Q 1=C h 2 ) P(h 2 ) P(Q 1 =C) = =
15 Computing P 1 (h i ) Consider P 1 (h 3 ). P 1 h 3 = P h 3 Q 1 = C) P h 3 Q 1 = C) = P Q 1=C h 3 ) P(h 3 ) P(Q 1 =C) = =
16 Computing P 1 (h i ) Consider P 1 (h 4 ). P 1 h 4 = P h 4 Q 1 = C) P h 4 Q 1 = C) = P Q 1=C h 4 ) P(h 4 ) P(Q 1 =C) = =
17 Computing P 1 (h i ) Consider P 1 (h 5 ). P 1 h 5 = P h 5 Q 1 = C) P h 5 Q 1 = C) = P Q 1=C h 5 ) P(h 5 ) P(Q 1 =C) = = 0 17
18 Updated Probabilities, after Q 1 = C P 0 (h 1 ) = 0.1 P 1 (h 1 ) = 0.2 : 100% cherry candies P 0 (h 2 ) = 0.2 P 1 (h 2 ) = 0.3 : h 2 : 75% cherry candies + 25% lime candies P 0 (h 3 ) = 0.4 P 1 (h 3 ) = 0.4 : h 3 : 50% cherry candies + 50% lime candies P 0 (h 4 ) = 0.2 P 1 (h 4 ) = 0.1 : h 4 : 25% cherry candies + 75% lime candies P 0 (h 5 ) = 0.1 P 1 (h 5 ) = 0.0 : h 5 : 100% lime candies Probabilities have changed for each bag, now that we have picked one candy and we have seen that it is a cherry candy. 18
19 Computing P 1 (Q t ) Now, consider P 1 (Q 2 = C). What does P 1 (Q 2 = C) mean? It is the probability that the second candy we pick out of our bag is a cherry candy, given our knowledge of the type of the first candy. To continue with our previous example, we assume that the first candy was cherry, so Q 1 = C. P 1 Q 2 = C = P Q 2 = C Q 1 = C) 19
20 Computing P 1 (Q t ) P 1 Q 2 = C = P Q 2 = C Q 1 = C) = P Q 1 = C h 1, Q 1 = C) P h 1 Q 1 = C) + P Q 1 = C h 2, Q 1 = C) P h 2 Q 1 = C) + P Q 1 = C h 3, Q 1 = C) P h 3 Q 1 = C) + P Q 1 = C h 4, Q 1 = C) P h 4 Q 1 = C) + P Q 1 = C h 5, Q 1 = C) P h 5 Q 1 = C) 20
21 Computing P 1 (Q t ) P 1 Q 2 = C = P Q 2 = C Q 1 = C) = P Q 1 = C h 1, Q 1 = C) P h 1 Q 1 = C) + P Q 1 = C h 2, Q 1 = C) P h 2 Q 1 = C) + P Q 1 = C h 3, Q 1 = C) P h 3 Q 1 = C) + P Q 1 = C h 4, Q 1 = C) P h 4 Q 1 = C) + P Q 1 = C h 5, Q 1 = C) P h 5 Q 1 = C) NOTE: Q 2 is conditionally independent of Q 1 given h i. If we know the type of bag, what we have already picked does not change our expectation of what we will pick next. Therefore, we can simplify P Q 1 = C h 1, Q 1 = C) as P Q 1 = C h 1 ). 21
22 Computing P 1 (Q t ) We simplify P Q 1 = C h 1, Q 1 = C) as P Q 1 = C h 1 ) P 1 Q 2 = C = P Q 2 = C Q 1 = C) = P Q 1 = C h 1 ) P h 1 Q 1 = C) + P Q 1 = C h 2 ) P h 2 Q 1 = C) + P Q 1 = C h 3 ) P h 3 Q 1 = C) + P Q 1 = C h 4 ) P h 4 Q 1 = C) + P Q 1 = C h 5 ) P h 5 Q 1 = C) 22
23 Computing P 1 (Q t ) We simplify P Q 1 = C h 1, Q 1 = C) as P Q 1 = C h 1 ) P 1 Q 2 = C = P Q 2 = C Q 1 = C) = P Q 1 = C h 1 ) P h 1 Q 1 = C) + P Q 1 = C h 2 ) P h 2 Q 1 = C) + P Q 1 = C h 3 ) P h 3 Q 1 = C) + P Q 1 = C h 4 ) P h 4 Q 1 = C) + P Q 1 = C h 5 ) P h 5 Q 1 = C) We now can plug in numbers everywhere: We know P Q 1 = C h i ). We computed P h i Q 1 = C) in previous slides, and we called it P 1 (h i ). 23
24 Computing P 1 (Q t ) P 1 Q 2 = C = P Q 2 = C Q 1 = C) = = 0.65 Notice the difference caused by the knowledge that Q 1 = C: P 0 Q 1 = C = 0.5 P 1 Q 2 = C =
25 Computing P 2 (h i ) Now, consider P 2 (h 1 ). What does P 2 (h 1 ) mean? We have defined P 2 (h 1 ) as the probability of h 1 after the first two observations. In our example, the first two observations were both of type cherry. Therefore: P 2 h 1 = P h 1 Q 1 = C, Q 2 = C) =??? 25
26 A Special Case of Bayes Rule The normal version of Bayes rule states that: P(A B) = P(B A) P(A) P B From the basic formula, we can derive a special case of Bayes rule, that we can apply if we also know some other fact F: P A B, F) = P B A, F) P A F) P B F Here, we want to compute P h 1 Q 1 = C, Q 2 = C). We will apply the special case of Bayes rule, with: h 1 as A. "Q 2 = C" as B. "Q 1 = C" as F. 26
27 Computing P 2 (h i ) P 2 h 1 = P h 1 Q 2 = C, Q 1 = C) = P Q 2 = C h 1, Q 1 = C) P h 1 Q 1 = C) P Q 2 = C Q 1 = C) These are all quantities we have computed before: 27
28 Computing P 2 (h i ) P 2 h 1 = P h 1 Q 2 = C, Q 1 = C) = P Q 2 = C h 1, Q 1 = C) P h 1 Q 1 = C) P Q 2 = C Q 1 = C) These are all quantities we have computed before: P Q 2 = C h 1, Q 1 = C) = P Q 2 = C h 1 ) = 1. P h 1 Q 1 = C) = P 1 h 1 = 0.2 P Q 2 = C Q 1 = C) = P 1 Q 2 = C =
29 Computing P 2 (h i ) P 2 h 1 = P h 1 Q 2 = C, Q 1 = C) = P Q 2 = C h 1, Q 1 = C) P h 1 Q 1 = C) P Q 2 = C Q 1 = C) = P Q 2 = C h 1 ) P 1 h 1 P 1 Q 2 = C =
30 Computing P 2 (h i ) P 2 h 2 = P h 2 Q 2 = C, Q 1 = C) = P Q 2 = C h 2, Q 1 = C) P h 2 Q 1 = C) P Q 2 = C Q 1 = C) = P Q 2 = C h 2 ) P 1 h 2 P 1 Q 2 = C =
31 Computing P 2 (h i ) P 2 h 3 = P h 3 Q 2 = C, Q 1 = C) = P Q 2 = C h 3, Q 1 = C) P h 3 Q 1 = C) P Q 2 = C Q 1 = C) = P Q 2 = C h 3 ) P 1 h 3 P 1 Q 2 = C =
32 Computing P 2 (h i ) P 2 h 4 = P h 4 Q 2 = C, Q 1 = C) = P Q 2 = C h 4, Q 1 = C) P h 4 Q 1 = C) P Q 2 = C Q 1 = C) = P Q 2 = C h 4 ) P 1 h 4 P 1 Q 2 = C =
33 Computing P 2 (h i ) P 2 h 5 = P h 5 Q 2 = C, Q 1 = C) = P Q 2 = C h 5, Q 1 = C) P h 5 Q 1 = C) P Q 2 = C Q 1 = C) = P Q 2 = C h 5 ) P 1 h 5 P 1 Q 2 = C = = 0 33
34 Updated Probabilities Probabilities of bags: Before any observations. After one observation (assuming the first candy is of type cherry). After two observations (assuming both candies are of type cherry). P 0 (h 1 ) = 0.1 P 1 (h 1 ) = 0.2 P 2 (h 1 ) = h 1 : 100% cherry candies P 0 (h 2 ) = 0.2 P 1 (h 2 ) = 0.3 P 2 (h 2 ) = h 2 : 75% cherry + 25% lime P 0 (h 3 ) = 0.4 P 1 (h 3 ) = 0.4 P 2 (h 3 ) = h 3 : 50% cherry + 50% lime P 0 (h 4 ) = 0.2 P 1 (h 4 ) = 0.1 P 2 (h 4 ) = h 4 : 25% cherry + 75% lime P 0 (h 5 ) = 0.1 P 1 (h 5 ) = 0.0 P 2 (h 5 ) = 0.0 h 5 : 100% lime candies 34
35 Computing P t (h i ) Let t be an integer between 1 and T: We have defined P t (h i ) = P(h i Q 1,, Q t ) To compute P t (h i ), we will use again the special case of Bayes rule that we used before: P A B, F) = P B A, F) P A F) P B F We will apply this formula, using: h i as A. Q t as B. "Q 1, Q 2,, Q t-1 " as F. 35
36 Computing P t (h i ) Let t be an integer between 1 and T: P t (h i ) = P(h i Q 1,, Q t ) = P(Q t h i, Q 1,, Q t-1 ) * P(h i Q 1,, Q t-1 ) P(Q t Q 1,, Q t-1 ) => => P t (h i ) = P(Q t h i ) * P t-1 (h i ) P t-1 (Q t ) 36
37 Computing P t+1 (Q t ) P t (Q t+1 ) = P(Q t+1 Q 1,, Q t ) = 5 Σ i=1 (P(Q t+1 h i ) P(h i Q 1,, Q t )) => 5 P t (Q t+1 ) = (P(Q t+1 h i ) P t (h i )) Σ i=1 37
38 Computing P t (h i ) (continued) The formula is recursive, as it requires knowing P t-1 (h i ). P t (h i ) = The base case is P 0 (h i ) = P(h i ). P(Q t h i ) * P t-1 (h i ) P t-1 (Q t ) To compute P t (h i ) we also need P t-1 (Q t ). We show how to compute that next. 38
39 Computing P t (h i ) and P t (Q t+1 ) Base case: t = 0. P 0 (h i ) = P(h i ), where P(h i ) is known. 5 Σ P 0 (Q 1 ) = ( P(Q 1 h i ) * P(h i ) ), where P(Q 1 h i ) is known. i=1 To compute P t (h i ) and P t (Q t+1 ): For j = 1,, t Compute P j (h i ) = P(Q j h i ) * P j-1 (h i ) P j-1 (Q j ) 5 Σ Compute P j (Q j+1 ) = ( P(Q j+1 h i ) * P j (h i )) i=1 39
40 Computing P t (h i ) and P t (Q t+1 ) Base case: t = 0. P 0 (h i ) = P(h i ), where P(h i ) is known. 5 Σ P 0 (Q 1 ) = ( P(Q 1 h i ) * P(h i ) ), where P(Q 1 h i ) is known. i=1 To compute P t (h i ) and P t (Q t+1 ): For j = 1,, t Compute P j (h i ) = known computed at previous round P(Q j h i ) * P j-1 (h i ) P j-1 (Q j ) computed at previous round 5 Σ Compute P j (Q j+1 ) = ( P(Q j+1 h i ) * P j (h i )) i=1 known computed at previous line 40
Statistical learning. Chapter 20, Sections 1 3 1
Statistical learning Chapter 20, Sections 1 3 Chapter 20, Sections 1 3 1 Outline Bayesian learning Maximum a posteriori and maximum likelihood learning Bayes net learning ML parameter learning with complete
More informationBayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction
15-0: Learning vs. Deduction Artificial Intelligence Programming Bayesian Learning Chris Brooks Department of Computer Science University of San Francisco So far, we ve seen two types of reasoning: Deductive
More informationLearning with Probabilities
Learning with Probabilities CS194-10 Fall 2011 Lecture 15 CS194-10 Fall 2011 Lecture 15 1 Outline Bayesian learning eliminates arbitrary loss functions and regularizers facilitates incorporation of prior
More informationStatistical learning. Chapter 20, Sections 1 3 1
Statistical learning Chapter 20, Sections 1 3 Chapter 20, Sections 1 3 1 Outline Bayesian learning Maximum a posteriori and maximum likelihood learning Bayes net learning ML parameter learning with complete
More informationBayesian Updating: Discrete Priors: Spring
Bayesian Updating: Discrete Priors: 18.05 Spring 2017 http://xkcd.com/1236/ Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial. 2. Treatment 2:
More informationFrom inductive inference to machine learning
From inductive inference to machine learning ADAPTED FROM AIMA SLIDES Russel&Norvig:Artificial Intelligence: a modern approach AIMA: Inductive inference AIMA: Inductive inference 1 Outline Bayesian inferences
More informationStatistical learning. Chapter 20, Sections 1 4 1
Statistical learning Chapter 20, Sections 1 4 Chapter 20, Sections 1 4 1 Outline Bayesian learning Maximum a posteriori and maximum likelihood learning Bayes net learning ML parameter learning with complete
More informationBayesian Updating: Discrete Priors: Spring
Bayesian Updating: Discrete Priors: 18.05 Spring 2017 http://xkcd.com/1236/ Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial. 2. Treatment 2:
More informationHidden Markov Models Part 2: Algorithms
Hidden Markov Models Part 2: Algorithms CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Hidden Markov Model An HMM consists of:
More informationMA : Introductory Probability
MA 320-001: Introductory Probability David Murrugarra Department of Mathematics, University of Kentucky http://www.math.uky.edu/~dmu228/ma320/ Spring 2017 David Murrugarra (University of Kentucky) MA 320:
More informationOutline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012
CSE 573: Artificial Intelligence Autumn 2012 Reasoning about Uncertainty & Hidden Markov Models Daniel Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 Outline
More informationBayesian networks. Chapter 14, Sections 1 4
Bayesian networks Chapter 14, Sections 1 4 Artificial Intelligence, spring 2013, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 14, Sections 1 4 1 Bayesian networks
More informationIntroduction to Bayesian Learning
Course Information Introduction Introduction to Bayesian Learning Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Apprendimento Automatico: Fondamenti - A.A. 2016/2017 Outline
More information15-381: Artificial Intelligence. Hidden Markov Models (HMMs)
15-381: Artificial Intelligence Hidden Markov Models (HMMs) What s wrong with Bayesian networks Bayesian networks are very useful for modeling joint distributions But they have their limitations: - Cannot
More informationMACHINE LEARNING. Probably Approximately Correct (PAC) Learning. Alessandro Moschitti
MACHINE LEARNING Probably Approximately Correct (PAC) Learning Alessandro Moschitti Department of Information Engineering and Computer Science University of Trento Email: moschitti@disi.unitn.it Objectives:
More informationLinear Classification. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington
Linear Classification CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Example of Linear Classification Red points: patterns belonging
More informationLinear Models for Regression
Linear Models for Regression CSE 4309 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 The Regression Problem Training data: A set of input-output
More informationFinal Examination CS 540-2: Introduction to Artificial Intelligence
Final Examination CS 540-2: Introduction to Artificial Intelligence May 7, 2017 LAST NAME: SOLUTIONS FIRST NAME: Problem Score Max Score 1 14 2 10 3 6 4 10 5 11 6 9 7 8 9 10 8 12 12 8 Total 100 1 of 11
More informationMAE 493G, CpE 493M, Mobile Robotics. 6. Basic Probability
MAE 493G, CpE 493M, Mobile Robotics 6. Basic Probability Instructor: Yu Gu, Fall 2013 Uncertainties in Robotics Robot environments are inherently unpredictable; Sensors and data acquisition systems are
More informationBayesian Classifiers and Probability Estimation. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington
Bayesian Classifiers and Probability Estimation Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington 1 Data Space Suppose that we have a classification problem The
More informationCSE 473: Artificial Intelligence
CSE 473: Artificial Intelligence Probability Steve Tanimoto University of Washington [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials
More informationStatistical Learning. Philipp Koehn. 10 November 2015
Statistical Learning Philipp Koehn 10 November 2015 Outline 1 Learning agents Inductive learning Decision tree learning Measuring learning performance Bayesian learning Maximum a posteriori and maximum
More informationCSE 312 Final Review: Section AA
CSE 312 TAs December 8, 2011 General Information General Information Comprehensive Midterm General Information Comprehensive Midterm Heavily weighted toward material after the midterm Pre-Midterm Material
More informationConstructive Artificial Intelligence
Probabilistic Inference Daniel Polani School of Computer Science University of Hertfordshire November 24, 2013 All rights reserved. Permission is granted to copy and distribute these slides in full or
More informationBayesian Networks BY: MOHAMAD ALSABBAGH
Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional
More informationLogistics. Naïve Bayes & Expectation Maximization. 573 Schedule. Coming Soon. Estimation Models. Topics
Logistics Naïve Bayes & Expectation Maximization CSE 7 eam Meetings Midterm Open book, notes Studying See AIMA exercises Daniel S. Weld Daniel S. Weld 7 Schedule Selected opics Coming Soon Selected opics
More informationConsider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes.
CMSC 310 Artificial Intelligence Probabilistic Reasoning and Bayesian Belief Networks Probabilities, Random Variables, Probability Distribution, Conditional Probability, Joint Distributions, Bayes Theorem
More informationIntroduction to Artificial Intelligence. Unit # 11
Introduction to Artificial Intelligence Unit # 11 1 Course Outline Overview of Artificial Intelligence State Space Representation Search Techniques Machine Learning Logic Probabilistic Reasoning/Bayesian
More informationAutonomous Mobile Robot Design
Autonomous Mobile Robot Design Topic: Particle Filter for Localization Dr. Kostas Alexis (CSE) These slides relied on the lectures from C. Stachniss, and the book Probabilistic Robotics from Thurn et al.
More informationBayesian RL Seminar. Chris Mansley September 9, 2008
Bayesian RL Seminar Chris Mansley September 9, 2008 Bayes Basic Probability One of the basic principles of probability theory, the chain rule, will allow us to derive most of the background material in
More informationBayesian Updating: Discrete Priors: Spring 2014
ian Updating: Discrete Priors: 18.05 Spring 2014 http://xkcd.com/1236/ January 1, 2017 1 / 16 Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial.
More informationHuman-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg
Temporal Reasoning Kai Arras, University of Freiburg 1 Temporal Reasoning Contents Introduction Temporal Reasoning Hidden Markov Models Linear Dynamical Systems (LDS) Kalman Filter 2 Temporal Reasoning
More informationCryptography: Joining the RSA Cryptosystem
Cryptography: Joining the RSA Cryptosystem Greg Plaxton Theory in Programming Practice, Fall 2005 Department of Computer Science University of Texas at Austin Joining the RSA Cryptosystem: Overview First,
More informationBayesian Networks. Machine Learning, Fall Slides based on material from the Russell and Norvig AI Book, Ch. 14
Bayesian Networks Machine Learning, Fall 2010 Slides based on material from the Russell and Norvig AI Book, Ch. 14 1 Administrativia Bayesian networks The inference problem: given a BN, how to make predictions
More informationCSE 473: Artificial Intelligence Probability Review à Markov Models. Outline
CSE 473: Artificial Intelligence Probability Review à Markov Models Daniel Weld University of Washington [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.
More informationProbabilistic Reasoning. Kee-Eung Kim KAIST Computer Science
Probabilistic Reasoning Kee-Eung Kim KAIST Computer Science Outline #1 Acting under uncertainty Probabilities Inference with Probabilities Independence and Bayes Rule Bayesian networks Inference in Bayesian
More informationNeural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington
Neural Networks CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Perceptrons x 0 = 1 x 1 x 2 z = h w T x Output: z x D A perceptron
More informationLearning Theory. Machine Learning CSE546 Carlos Guestrin University of Washington. November 25, Carlos Guestrin
Learning Theory Machine Learning CSE546 Carlos Guestrin University of Washington November 25, 2013 Carlos Guestrin 2005-2013 1 What now n We have explored many ways of learning from data n But How good
More informationCSEP 573: Artificial Intelligence
CSEP 573: Artificial Intelligence Bayesian Networks: Inference Ali Farhadi Many slides over the course adapted from either Luke Zettlemoyer, Pieter Abbeel, Dan Klein, Stuart Russell or Andrew Moore 1 Outline
More informationCS 343: Artificial Intelligence
CS 343: Artificial Intelligence Probability Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188
More informationInfinite Continued Fractions
Infinite Continued Fractions 8-5-200 The value of an infinite continued fraction [a 0 ; a, a 2, ] is lim c k, where c k is the k-th convergent k If [a 0 ; a, a 2, ] is an infinite continued fraction with
More informationlast two digits of your SID
Announcements Midterm: Wednesday 7pm-9pm See midterm prep page (posted on Piazza, inst.eecs page) Four rooms; your room determined by last two digits of your SID: 00-32: Dwinelle 155 33-45: Genetics and
More informationText Categorization CSE 454. (Based on slides by Dan Weld, Tom Mitchell, and others)
Text Categorization CSE 454 (Based on slides by Dan Weld, Tom Mitchell, and others) 1 Given: Categorization A description of an instance, x X, where X is the instance language or instance space. A fixed
More informationLast Time. Today. Bayesian Learning. The Distributions We Love. CSE 446 Gaussian Naïve Bayes & Logistic Regression
CSE 446 Gaussian Naïve Bayes & Logistic Regression Winter 22 Dan Weld Learning Gaussians Naïve Bayes Last Time Gaussians Naïve Bayes Logistic Regression Today Some slides from Carlos Guestrin, Luke Zettlemoyer
More informationLearning and Neural Networks
Artificial Intelligence Learning and Neural Networks Readings: Chapter 19 & 20.5 of Russell & Norvig Example: A Feed-forward Network w 13 I 1 H 3 w 35 w 14 O 5 I 2 w 23 w 24 H 4 w 45 a 5 = g 5 (W 3,5 a
More informationRecall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem
Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)
More informationPriors, Total Probability, Expectation, Multiple Trials
Priors, Total Probability, Expectation, Multiple Trials CS 2800: Discrete Structures, Fall 2014 Sid Chaudhuri Bayes' Theorem Given: prior probabilities of hypotheses, and the probability that each hypothesis
More informationIntroduction to Artificial Intelligence Belief networks
Introduction to Artificial Intelligence Belief networks Chapter 15.1 2 Dieter Fox Based on AIMA Slides c S. Russell and P. Norvig, 1998 Chapter 15.1 2 0-0 Outline Bayesian networks: syntax and semantics
More informationPreliminary check: are the terms that we are adding up go to zero or not? If not, proceed! If the terms a n are going to zero, pick another test.
Throughout these templates, let series. be a series. We hope to determine the convergence of this Divergence Test: If lim is not zero or does not exist, then the series diverges. Preliminary check: are
More informationCSE 473: Artificial Intelligence Autumn Topics
CSE 473: Artificial Intelligence Autumn 2014 Bayesian Networks Learning II Dan Weld Slides adapted from Jack Breese, Dan Klein, Daphne Koller, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 473 Topics
More informationMATH 240. Chapter 8 Outlines of Hypothesis Tests
MATH 4 Chapter 8 Outlines of Hypothesis Tests Test for Population Proportion p Specify the null and alternative hypotheses, ie, choose one of the three, where p is some specified number: () H : p H : p
More informationGrundlagen der Künstlichen Intelligenz
Grundlagen der Künstlichen Intelligenz Uncertainty & Probabilities & Bandits Daniel Hennes 16.11.2017 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Uncertainty Probability
More informationArtificial neural networks
Artificial neural networks Chapter 8, Section 7 Artificial Intelligence, spring 203, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 8, Section 7 Outline Brains Neural
More informationAdam Blank Spring 2017 CSE 311. Foundations of Computing I
Adam Blank Spring 2017 CSE 311 Foundations of Computing I Pre-Lecture Problem Suppose that p, and p (q r) are true. Is q true? Can you prove it with equivalences? CSE 311: Foundations of Computing Lecture
More informationCOMP9414: Artificial Intelligence Reasoning Under Uncertainty
COMP9414, Monday 16 April, 2012 Reasoning Under Uncertainty 2 COMP9414: Artificial Intelligence Reasoning Under Uncertainty Overview Problems with Logical Approach What Do the Numbers Mean? Wayne Wobcke
More informationProbabilistic Robotics
University of Rome La Sapienza Master in Artificial Intelligence and Robotics Probabilistic Robotics Prof. Giorgio Grisetti Course web site: http://www.dis.uniroma1.it/~grisetti/teaching/probabilistic_ro
More informationUncertainty. Chapter 13, Sections 1 6
Uncertainty Chapter 13, Sections 1 6 Artificial Intelligence, spring 2013, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 13, Sections 1 6 1 Outline Uncertainty Probability
More informationOur Status in CSE 5522
Our Status in CSE 5522 We re done with Part I Search and Planning! Part II: Probabilistic Reasoning Diagnosis Speech recognition Tracking objects Robot mapping Genetics Error correcting codes lots more!
More informationLearning Objectives. c D. Poole and A. Mackworth 2010 Artificial Intelligence, Lecture 7.2, Page 1
Learning Objectives At the end of the class you should be able to: identify a supervised learning problem characterize how the prediction is a function of the error measure avoid mixing the training and
More informationProbability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides
Probability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides slide 1 Inference with Bayes rule: Example In a bag there are two envelopes one has a red ball (worth $100) and a black ball one
More informationBayesian Methods: Naïve Bayes
Bayesian Methods: aïve Bayes icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Last Time Parameter learning Learning the parameter of a simple coin flipping model Prior
More informationIntroduction to Artificial Intelligence Midterm 2. CS 188 Spring You have approximately 2 hours and 50 minutes.
CS 188 Spring 2014 Introduction to Artificial Intelligence Midterm 2 You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your two-page crib sheet. Mark your answers
More informationToday. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use?
Today Statistical Learning Parameter Estimation: Maximum Likelihood (ML) Maximum A Posteriori (MAP) Bayesian Continuous case Learning Parameters for a Bayesian Network Naive Bayes Maximum Likelihood estimates
More informationDiscrete Structures - CM0246 Cardinality
Discrete Structures - CM0246 Cardinality Andrés Sicard-Ramírez Universidad EAFIT Semester 2014-2 Cardinality Definition (Cardinality (finite sets)) Let A be a set. The number of (distinct) elements in
More informationCS188 Outline. We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning. Part III: Machine Learning
CS188 Outline We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning Diagnosis Speech recognition Tracking objects Robot mapping Genetics Error correcting codes lots more! Part III:
More informationLecture Slides. Elementary Statistics. by Mario F. Triola. and the Triola Statistics Series
Lecture Slides Elementary Statistics Tenth Edition and the Triola Statistics Series by Mario F. Triola Slide 1 Chapter 9 Inferences from Two Samples 9-1 Overview 9-2 Inferences About Two Proportions 9-3
More informationHidden Markov Models Part 1: Introduction
Hidden Markov Models Part 1: Introduction CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Modeling Sequential Data Suppose that
More informationOutline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car
CSE 573: Artificial Intelligence Autumn 2012 Bayesian Networks Dan Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer Outline Probabilistic models (and inference)
More informationEQ: What are limits, and how do we find them? Finite limits as x ± Horizontal Asymptote. Example Horizontal Asymptote
Finite limits as x ± The symbol for infinity ( ) does not represent a real number. We use to describe the behavior of a function when the values in its domain or range outgrow all finite bounds. For example,
More informationFinal Exam December 12, 2017
Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes
More informationBiointelligence Lab School of Computer Sci. & Eng. Seoul National University
Artificial Intelligence Chater 19 easoning with Uncertain Information Biointelligence Lab School of Comuter Sci. & Eng. Seoul National University Outline l eview of Probability Theory l Probabilistic Inference
More informationCSE 473: Artificial Intelligence Autumn 2011
CSE 473: Artificial Intelligence Autumn 2011 Bayesian Networks Luke Zettlemoyer Many slides over the course adapted from either Dan Klein, Stuart Russell or Andrew Moore 1 Outline Probabilistic models
More informationCS 343: Artificial Intelligence
CS 343: Artificial Intelligence Bayes Nets: Sampling Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.
More informationComputational Cognitive Science
Computational Cognitive Science Lecture 8: Frank Keller School of Informatics University of Edinburgh keller@inf.ed.ac.uk Based on slides by Sharon Goldwater October 14, 2016 Frank Keller Computational
More informationProbability. CS 3793/5233 Artificial Intelligence Probability 1
CS 3793/5233 Artificial Intelligence 1 Motivation Motivation Random Variables Semantics Dice Example Joint Dist. Ex. Axioms Agents don t have complete knowledge about the world. Agents need to make decisions
More informationParametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a
Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Some slides are due to Christopher Bishop Limitations of K-means Hard assignments of data points to clusters small shift of a
More informationMODULE 10 Bayes Classifier LESSON 20
MODULE 10 Bayes Classifier LESSON 20 Bayesian Belief Networks Keywords: Belief, Directed Acyclic Graph, Graphical Model 1 Bayesian Belief Network A Bayesian network is a graphical model of a situation
More informationModel Averaging (Bayesian Learning)
Model Averaging (Bayesian Learning) We want to predict the output Y of a new case that has input X = x given the training examples e: p(y x e) = m M P(Y m x e) = m M P(Y m x e)p(m x e) = m M P(Y m x)p(m
More informationINDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. NPTEL National Programme on Technology Enhanced Learning. Probability Methods in Civil Engineering
INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR NPTEL National Programme on Technology Enhanced Learning Probability Methods in Civil Engineering Prof. Rajib Maity Department of Civil Engineering IIT Kharagpur
More informationBayes Theorem (10B) Young Won Lim 6/3/17
Bayes Theorem (10B) Copyright (c) 2017 Young W. Lim. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later
More informationSec$on Summary. Mathematical Proofs Forms of Theorems Trivial & Vacuous Proofs Direct Proofs Indirect Proofs
Section 1.7 Sec$on Summary Mathematical Proofs Forms of Theorems Trivial & Vacuous Proofs Direct Proofs Indirect Proofs Proof of the Contrapositive Proof by Contradiction 2 Proofs of Mathema$cal Statements
More informationBayes Theorem (4A) Young Won Lim 3/5/18
Bayes Theorem (4A) Copyright (c) 2017-2018 Young W. Lim. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any
More informationCS 231A Section 1: Linear Algebra & Probability Review
CS 231A Section 1: Linear Algebra & Probability Review 1 Topics Support Vector Machines Boosting Viola-Jones face detector Linear Algebra Review Notation Operations & Properties Matrix Calculus Probability
More informationFinal Exam December 12, 2017
Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes
More informationCS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang
CS 231A Section 1: Linear Algebra & Probability Review Kevin Tang Kevin Tang Section 1-1 9/30/2011 Topics Support Vector Machines Boosting Viola Jones face detector Linear Algebra Review Notation Operations
More informationComputational Perception. Bayesian Inference
Computational Perception 15-485/785 January 24, 2008 Bayesian Inference The process of probabilistic inference 1. define model of problem 2. derive posterior distributions and estimators 3. estimate parameters
More informationCS 188: Artificial Intelligence. Our Status in CS188
CS 188: Artificial Intelligence Probability Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. 1 Our Status in CS188 We re done with Part I Search and Planning! Part II: Probabilistic Reasoning
More informationKalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein
Kalman filtering and friends: Inference in time series models Herke van Hoof slides mostly by Michael Rubinstein Problem overview Goal Estimate most probable state at time k using measurement up to time
More informationBe able to define the following terms and answer basic questions about them:
CS440/ECE448 Section Q Fall 2017 Final Review Be able to define the following terms and answer basic questions about them: Probability o Random variables, axioms of probability o Joint, marginal, conditional
More informationAlgorithm Analysis Recurrence Relation. Chung-Ang University, Jaesung Lee
Algorithm Analysis Recurrence Relation Chung-Ang University, Jaesung Lee Recursion 2 Recursion 3 Recursion in Real-world Fibonacci sequence = + Initial conditions: = 0 and = 1. = + = + = + 0, 1, 1, 2,
More informationProbability Basics. Robot Image Credit: Viktoriya Sukhanova 123RF.com
Probability Basics These slides were assembled by Eric Eaton, with grateful acknowledgement of the many others who made their course materials freely available online. Feel free to reuse or adapt these
More informationAnnouncements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation
CS 188: Artificial Intelligence Spring 2010 Lecture 15: Bayes Nets II Independence 3/9/2010 Pieter Abbeel UC Berkeley Many slides over the course adapted from Dan Klein, Stuart Russell, Andrew Moore Current
More informationBayes Nets III: Inference
1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy
More informationBayesian Learning. Chapter 6: Bayesian Learning. Bayes Theorem. Roles for Bayesian Methods. CS 536: Machine Learning Littman (Wu, TA)
Bayesian Learning Chapter 6: Bayesian Learning CS 536: Machine Learning Littan (Wu, TA) [Read Ch. 6, except 6.3] [Suggested exercises: 6.1, 6.2, 6.6] Bayes Theore MAP, ML hypotheses MAP learners Miniu
More informationIntroduction to Spring 2006 Artificial Intelligence Practice Final
NAME: SID#: Login: Sec: 1 CS 188 Introduction to Spring 2006 Artificial Intelligence Practice Final You have 180 minutes. The exam is open-book, open-notes, no electronics other than basic calculators.
More informationAnnouncements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.
Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides
More informationMathematical Induction. Defining Functions. Overview. Notation for recursive functions. Base case Sn(0) = 0 S(n) = S(n 1) + n for n > 0
Readings on induction. Mathematical Induction (a) Weiss, Sec. 7.2, page 233 (b) Course slides f lecture and notes recitation. Every criticism from a good man is of value to me. What you hint at generally
More informationContinuous Data with Continuous Priors Class 14, Jeremy Orloff and Jonathan Bloom
1 Learning Goals Continuous Data with Continuous Priors Class 14, 18.0 Jeremy Orloff and Jonathan Bloom 1. Be able to construct a ian update table for continuous hypotheses and continuous data. 2. Be able
More informationCSE 21 Practice Exam for Midterm 2 Fall 2017
CSE 1 Practice Exam for Midterm Fall 017 These practice problems should help prepare you for the second midterm, which is on monday, November 11 This is longer than the actual exam will be, but good practice
More informationSample solutions to Homework 4, Information-Theoretic Modeling (Fall 2014)
Sample solutions to Homework 4, Information-Theoretic Modeling (Fall 204) Jussi Määttä October 2, 204 Question [First, note that we use the symbol! as an end-of-message symbol. When we see it, we know
More information