Computing Posterior Probabilities. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington

Size: px
Start display at page:

Download "Computing Posterior Probabilities. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington"

Transcription

1 Computing Posterior Probabilities Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington 1

2 Overview of Candy Bag Example As described in Russell and Norvig, for Chapter 20 of the 2 nd edition: Five kinds of bags of candies. 10% are h 1 : 100% cherry candies 20% are h 2 : 75% cherry candies + 25% lime candies 40% are h 3 : 50% cherry candies + 50% lime candies 20% are h 4 : 25% cherry candies + 75% lime candies 10% are h 5 : 100% lime candies Each bag has an infinite number of candies. This way, the ratio of candy types inside a bag does not change as we pick candies out of the bag. We have a bag, and we are picking candies out of it. Based on the types of candies we are picking, we want to figure out what type of bag we have. 2

3 Hypotheses and Prior Probabilities Five kinds of bags of candies. 10% are h 1 : 100% cherry candies 20% are h 2 : 75% cherry candies + 25% lime candies 40% are h 3 : 50% cherry candies + 50% lime candies 20% are h 4 : 25% cherry candies + 75% lime candies 10% are h 5 : 100% lime candies Each h i is called a hypothesis. The initial probability that is given for each hypothesis is called the prior probability for that hypothesis. It is called prior because it is the probability we have before we have made any observations. 3

4 Observations and Posteriors Out of our bag, we pick T candies, whose types are: Q 1, Q 2,, Q T. Each Q j is equal to either C (cherry) or L ( lime ). These Q j s are called the observations. Based on our observations, we want to answer two types of questions: What is P(h i Q 1,, Q t )? Probability of hypothesis i after t observations. This is called the posterior probability of h i. What is P(Q t+1 = C Q 1,, Q t )? Similarly, what is P(Q t+1 = L Q 1,, Q t ) Probability of observation t+1 after t observations. 4

5 Simplifying notation Define: P t (h i ) = P(h i Q 1,, Q t ) P t (Q t+1 = C) = P(Q t+1 = C Q 1,, Q t )? Special case: t = 0 (no observations): P 0 (h i ) = P(h i ) P 0 (h i ) is the prior probability of h i P 0 (Q 1 = C) = P(Q 1 = C) P 0 (Q 1 = C) is the probability that the first observation is equal to C. 5

6 Questions We Want to Answer, Revisited Using the simplified notation of the previous slide: What is P t (h i )? Posterior probability of hypothesis i after t observations. What is P t (Q t+1 = C)? Similarly, what is P t (Q t+1 = L) Probability of observation t+1 after t observations. 6

7 Computing P 0 (Q t ) As an example: Consider P 0 (Q 1 = C). What does P 0 (Q 1 = C) mean? 7

8 Computing P 0 (Q t ) As an example: Consider P 0 (Q 1 = C). What does P 0 (Q 1 = C) mean? It is the probability that the first candy we pick out of our bag is a cherry candy. P 0 Q 1 = C = P Q 1 = C h 1 ) P(h 1 ) + P Q 1 = C h 2 ) P(h 2 ) + P Q 1 = C h 3 ) P(h 3 ) + P Q 1 = C h 4 ) P(h 4 ) + P Q 1 = C h 5 ) P(h 5 ) 8

9 Computing P 0 (Q t ) As an example: Consider P 0 (Q 1 = C). What does P 0 (Q 1 = C) mean? It is the probability that the first candy we pick out of our bag is a cherry candy. P 0 Q 1 = C = = 0.5 9

10 Computing P 1 (h i ) As an example: Consider P 1 (h 1 ). What does P 1 (h 1 )mean? P 1 h 1 = P h 1 Q 1 = C) P 1 h 1 is the probability that our bag is of type h 1, if the first candy we pick out of our bag is a cherry candy. 10

11 Computing P 1 (h i ) As an example: Consider P 1 (h 1 ). P 1 h 1 = P h 1 Q 1 = C) P h 1 Q 1 = C) = 11

12 Computing P 1 (h i ) As an example: Consider P 1 (h 1 ). P 1 h 1 = P h 1 Q 1 = C) P h 1 Q 1 = C) = P Q 1=C h 1 ) P(h 1 ) P(Q 1 =C) 12

13 Computing P 1 (h i ) As an example: Consider P 1 (h 1 ). P 1 h 1 = P h 1 Q 1 = C) P h 1 Q 1 = C) = P Q 1=C h 1 ) P(h 1 ) P(Q 1 =C) = =

14 Computing P 1 (h i ) Consider P 1 (h 2 ). P 1 h 2 = P h 2 Q 1 = C) P h 2 Q 1 = C) = P Q 1=C h 2 ) P(h 2 ) P(Q 1 =C) = =

15 Computing P 1 (h i ) Consider P 1 (h 3 ). P 1 h 3 = P h 3 Q 1 = C) P h 3 Q 1 = C) = P Q 1=C h 3 ) P(h 3 ) P(Q 1 =C) = =

16 Computing P 1 (h i ) Consider P 1 (h 4 ). P 1 h 4 = P h 4 Q 1 = C) P h 4 Q 1 = C) = P Q 1=C h 4 ) P(h 4 ) P(Q 1 =C) = =

17 Computing P 1 (h i ) Consider P 1 (h 5 ). P 1 h 5 = P h 5 Q 1 = C) P h 5 Q 1 = C) = P Q 1=C h 5 ) P(h 5 ) P(Q 1 =C) = = 0 17

18 Updated Probabilities, after Q 1 = C P 0 (h 1 ) = 0.1 P 1 (h 1 ) = 0.2 : 100% cherry candies P 0 (h 2 ) = 0.2 P 1 (h 2 ) = 0.3 : h 2 : 75% cherry candies + 25% lime candies P 0 (h 3 ) = 0.4 P 1 (h 3 ) = 0.4 : h 3 : 50% cherry candies + 50% lime candies P 0 (h 4 ) = 0.2 P 1 (h 4 ) = 0.1 : h 4 : 25% cherry candies + 75% lime candies P 0 (h 5 ) = 0.1 P 1 (h 5 ) = 0.0 : h 5 : 100% lime candies Probabilities have changed for each bag, now that we have picked one candy and we have seen that it is a cherry candy. 18

19 Computing P 1 (Q t ) Now, consider P 1 (Q 2 = C). What does P 1 (Q 2 = C) mean? It is the probability that the second candy we pick out of our bag is a cherry candy, given our knowledge of the type of the first candy. To continue with our previous example, we assume that the first candy was cherry, so Q 1 = C. P 1 Q 2 = C = P Q 2 = C Q 1 = C) 19

20 Computing P 1 (Q t ) P 1 Q 2 = C = P Q 2 = C Q 1 = C) = P Q 1 = C h 1, Q 1 = C) P h 1 Q 1 = C) + P Q 1 = C h 2, Q 1 = C) P h 2 Q 1 = C) + P Q 1 = C h 3, Q 1 = C) P h 3 Q 1 = C) + P Q 1 = C h 4, Q 1 = C) P h 4 Q 1 = C) + P Q 1 = C h 5, Q 1 = C) P h 5 Q 1 = C) 20

21 Computing P 1 (Q t ) P 1 Q 2 = C = P Q 2 = C Q 1 = C) = P Q 1 = C h 1, Q 1 = C) P h 1 Q 1 = C) + P Q 1 = C h 2, Q 1 = C) P h 2 Q 1 = C) + P Q 1 = C h 3, Q 1 = C) P h 3 Q 1 = C) + P Q 1 = C h 4, Q 1 = C) P h 4 Q 1 = C) + P Q 1 = C h 5, Q 1 = C) P h 5 Q 1 = C) NOTE: Q 2 is conditionally independent of Q 1 given h i. If we know the type of bag, what we have already picked does not change our expectation of what we will pick next. Therefore, we can simplify P Q 1 = C h 1, Q 1 = C) as P Q 1 = C h 1 ). 21

22 Computing P 1 (Q t ) We simplify P Q 1 = C h 1, Q 1 = C) as P Q 1 = C h 1 ) P 1 Q 2 = C = P Q 2 = C Q 1 = C) = P Q 1 = C h 1 ) P h 1 Q 1 = C) + P Q 1 = C h 2 ) P h 2 Q 1 = C) + P Q 1 = C h 3 ) P h 3 Q 1 = C) + P Q 1 = C h 4 ) P h 4 Q 1 = C) + P Q 1 = C h 5 ) P h 5 Q 1 = C) 22

23 Computing P 1 (Q t ) We simplify P Q 1 = C h 1, Q 1 = C) as P Q 1 = C h 1 ) P 1 Q 2 = C = P Q 2 = C Q 1 = C) = P Q 1 = C h 1 ) P h 1 Q 1 = C) + P Q 1 = C h 2 ) P h 2 Q 1 = C) + P Q 1 = C h 3 ) P h 3 Q 1 = C) + P Q 1 = C h 4 ) P h 4 Q 1 = C) + P Q 1 = C h 5 ) P h 5 Q 1 = C) We now can plug in numbers everywhere: We know P Q 1 = C h i ). We computed P h i Q 1 = C) in previous slides, and we called it P 1 (h i ). 23

24 Computing P 1 (Q t ) P 1 Q 2 = C = P Q 2 = C Q 1 = C) = = 0.65 Notice the difference caused by the knowledge that Q 1 = C: P 0 Q 1 = C = 0.5 P 1 Q 2 = C =

25 Computing P 2 (h i ) Now, consider P 2 (h 1 ). What does P 2 (h 1 ) mean? We have defined P 2 (h 1 ) as the probability of h 1 after the first two observations. In our example, the first two observations were both of type cherry. Therefore: P 2 h 1 = P h 1 Q 1 = C, Q 2 = C) =??? 25

26 A Special Case of Bayes Rule The normal version of Bayes rule states that: P(A B) = P(B A) P(A) P B From the basic formula, we can derive a special case of Bayes rule, that we can apply if we also know some other fact F: P A B, F) = P B A, F) P A F) P B F Here, we want to compute P h 1 Q 1 = C, Q 2 = C). We will apply the special case of Bayes rule, with: h 1 as A. "Q 2 = C" as B. "Q 1 = C" as F. 26

27 Computing P 2 (h i ) P 2 h 1 = P h 1 Q 2 = C, Q 1 = C) = P Q 2 = C h 1, Q 1 = C) P h 1 Q 1 = C) P Q 2 = C Q 1 = C) These are all quantities we have computed before: 27

28 Computing P 2 (h i ) P 2 h 1 = P h 1 Q 2 = C, Q 1 = C) = P Q 2 = C h 1, Q 1 = C) P h 1 Q 1 = C) P Q 2 = C Q 1 = C) These are all quantities we have computed before: P Q 2 = C h 1, Q 1 = C) = P Q 2 = C h 1 ) = 1. P h 1 Q 1 = C) = P 1 h 1 = 0.2 P Q 2 = C Q 1 = C) = P 1 Q 2 = C =

29 Computing P 2 (h i ) P 2 h 1 = P h 1 Q 2 = C, Q 1 = C) = P Q 2 = C h 1, Q 1 = C) P h 1 Q 1 = C) P Q 2 = C Q 1 = C) = P Q 2 = C h 1 ) P 1 h 1 P 1 Q 2 = C =

30 Computing P 2 (h i ) P 2 h 2 = P h 2 Q 2 = C, Q 1 = C) = P Q 2 = C h 2, Q 1 = C) P h 2 Q 1 = C) P Q 2 = C Q 1 = C) = P Q 2 = C h 2 ) P 1 h 2 P 1 Q 2 = C =

31 Computing P 2 (h i ) P 2 h 3 = P h 3 Q 2 = C, Q 1 = C) = P Q 2 = C h 3, Q 1 = C) P h 3 Q 1 = C) P Q 2 = C Q 1 = C) = P Q 2 = C h 3 ) P 1 h 3 P 1 Q 2 = C =

32 Computing P 2 (h i ) P 2 h 4 = P h 4 Q 2 = C, Q 1 = C) = P Q 2 = C h 4, Q 1 = C) P h 4 Q 1 = C) P Q 2 = C Q 1 = C) = P Q 2 = C h 4 ) P 1 h 4 P 1 Q 2 = C =

33 Computing P 2 (h i ) P 2 h 5 = P h 5 Q 2 = C, Q 1 = C) = P Q 2 = C h 5, Q 1 = C) P h 5 Q 1 = C) P Q 2 = C Q 1 = C) = P Q 2 = C h 5 ) P 1 h 5 P 1 Q 2 = C = = 0 33

34 Updated Probabilities Probabilities of bags: Before any observations. After one observation (assuming the first candy is of type cherry). After two observations (assuming both candies are of type cherry). P 0 (h 1 ) = 0.1 P 1 (h 1 ) = 0.2 P 2 (h 1 ) = h 1 : 100% cherry candies P 0 (h 2 ) = 0.2 P 1 (h 2 ) = 0.3 P 2 (h 2 ) = h 2 : 75% cherry + 25% lime P 0 (h 3 ) = 0.4 P 1 (h 3 ) = 0.4 P 2 (h 3 ) = h 3 : 50% cherry + 50% lime P 0 (h 4 ) = 0.2 P 1 (h 4 ) = 0.1 P 2 (h 4 ) = h 4 : 25% cherry + 75% lime P 0 (h 5 ) = 0.1 P 1 (h 5 ) = 0.0 P 2 (h 5 ) = 0.0 h 5 : 100% lime candies 34

35 Computing P t (h i ) Let t be an integer between 1 and T: We have defined P t (h i ) = P(h i Q 1,, Q t ) To compute P t (h i ), we will use again the special case of Bayes rule that we used before: P A B, F) = P B A, F) P A F) P B F We will apply this formula, using: h i as A. Q t as B. "Q 1, Q 2,, Q t-1 " as F. 35

36 Computing P t (h i ) Let t be an integer between 1 and T: P t (h i ) = P(h i Q 1,, Q t ) = P(Q t h i, Q 1,, Q t-1 ) * P(h i Q 1,, Q t-1 ) P(Q t Q 1,, Q t-1 ) => => P t (h i ) = P(Q t h i ) * P t-1 (h i ) P t-1 (Q t ) 36

37 Computing P t+1 (Q t ) P t (Q t+1 ) = P(Q t+1 Q 1,, Q t ) = 5 Σ i=1 (P(Q t+1 h i ) P(h i Q 1,, Q t )) => 5 P t (Q t+1 ) = (P(Q t+1 h i ) P t (h i )) Σ i=1 37

38 Computing P t (h i ) (continued) The formula is recursive, as it requires knowing P t-1 (h i ). P t (h i ) = The base case is P 0 (h i ) = P(h i ). P(Q t h i ) * P t-1 (h i ) P t-1 (Q t ) To compute P t (h i ) we also need P t-1 (Q t ). We show how to compute that next. 38

39 Computing P t (h i ) and P t (Q t+1 ) Base case: t = 0. P 0 (h i ) = P(h i ), where P(h i ) is known. 5 Σ P 0 (Q 1 ) = ( P(Q 1 h i ) * P(h i ) ), where P(Q 1 h i ) is known. i=1 To compute P t (h i ) and P t (Q t+1 ): For j = 1,, t Compute P j (h i ) = P(Q j h i ) * P j-1 (h i ) P j-1 (Q j ) 5 Σ Compute P j (Q j+1 ) = ( P(Q j+1 h i ) * P j (h i )) i=1 39

40 Computing P t (h i ) and P t (Q t+1 ) Base case: t = 0. P 0 (h i ) = P(h i ), where P(h i ) is known. 5 Σ P 0 (Q 1 ) = ( P(Q 1 h i ) * P(h i ) ), where P(Q 1 h i ) is known. i=1 To compute P t (h i ) and P t (Q t+1 ): For j = 1,, t Compute P j (h i ) = known computed at previous round P(Q j h i ) * P j-1 (h i ) P j-1 (Q j ) computed at previous round 5 Σ Compute P j (Q j+1 ) = ( P(Q j+1 h i ) * P j (h i )) i=1 known computed at previous line 40

Statistical learning. Chapter 20, Sections 1 3 1

Statistical learning. Chapter 20, Sections 1 3 1 Statistical learning Chapter 20, Sections 1 3 Chapter 20, Sections 1 3 1 Outline Bayesian learning Maximum a posteriori and maximum likelihood learning Bayes net learning ML parameter learning with complete

More information

Bayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction

Bayesian Learning. Artificial Intelligence Programming. 15-0: Learning vs. Deduction 15-0: Learning vs. Deduction Artificial Intelligence Programming Bayesian Learning Chris Brooks Department of Computer Science University of San Francisco So far, we ve seen two types of reasoning: Deductive

More information

Learning with Probabilities

Learning with Probabilities Learning with Probabilities CS194-10 Fall 2011 Lecture 15 CS194-10 Fall 2011 Lecture 15 1 Outline Bayesian learning eliminates arbitrary loss functions and regularizers facilitates incorporation of prior

More information

Statistical learning. Chapter 20, Sections 1 3 1

Statistical learning. Chapter 20, Sections 1 3 1 Statistical learning Chapter 20, Sections 1 3 Chapter 20, Sections 1 3 1 Outline Bayesian learning Maximum a posteriori and maximum likelihood learning Bayes net learning ML parameter learning with complete

More information

Bayesian Updating: Discrete Priors: Spring

Bayesian Updating: Discrete Priors: Spring Bayesian Updating: Discrete Priors: 18.05 Spring 2017 http://xkcd.com/1236/ Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial. 2. Treatment 2:

More information

From inductive inference to machine learning

From inductive inference to machine learning From inductive inference to machine learning ADAPTED FROM AIMA SLIDES Russel&Norvig:Artificial Intelligence: a modern approach AIMA: Inductive inference AIMA: Inductive inference 1 Outline Bayesian inferences

More information

Statistical learning. Chapter 20, Sections 1 4 1

Statistical learning. Chapter 20, Sections 1 4 1 Statistical learning Chapter 20, Sections 1 4 Chapter 20, Sections 1 4 1 Outline Bayesian learning Maximum a posteriori and maximum likelihood learning Bayes net learning ML parameter learning with complete

More information

Bayesian Updating: Discrete Priors: Spring

Bayesian Updating: Discrete Priors: Spring Bayesian Updating: Discrete Priors: 18.05 Spring 2017 http://xkcd.com/1236/ Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial. 2. Treatment 2:

More information

Hidden Markov Models Part 2: Algorithms

Hidden Markov Models Part 2: Algorithms Hidden Markov Models Part 2: Algorithms CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Hidden Markov Model An HMM consists of:

More information

MA : Introductory Probability

MA : Introductory Probability MA 320-001: Introductory Probability David Murrugarra Department of Mathematics, University of Kentucky http://www.math.uky.edu/~dmu228/ma320/ Spring 2017 David Murrugarra (University of Kentucky) MA 320:

More information

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012 CSE 573: Artificial Intelligence Autumn 2012 Reasoning about Uncertainty & Hidden Markov Models Daniel Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 Outline

More information

Bayesian networks. Chapter 14, Sections 1 4

Bayesian networks. Chapter 14, Sections 1 4 Bayesian networks Chapter 14, Sections 1 4 Artificial Intelligence, spring 2013, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 14, Sections 1 4 1 Bayesian networks

More information

Introduction to Bayesian Learning

Introduction to Bayesian Learning Course Information Introduction Introduction to Bayesian Learning Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Apprendimento Automatico: Fondamenti - A.A. 2016/2017 Outline

More information

15-381: Artificial Intelligence. Hidden Markov Models (HMMs)

15-381: Artificial Intelligence. Hidden Markov Models (HMMs) 15-381: Artificial Intelligence Hidden Markov Models (HMMs) What s wrong with Bayesian networks Bayesian networks are very useful for modeling joint distributions But they have their limitations: - Cannot

More information

MACHINE LEARNING. Probably Approximately Correct (PAC) Learning. Alessandro Moschitti

MACHINE LEARNING. Probably Approximately Correct (PAC) Learning. Alessandro Moschitti MACHINE LEARNING Probably Approximately Correct (PAC) Learning Alessandro Moschitti Department of Information Engineering and Computer Science University of Trento Email: moschitti@disi.unitn.it Objectives:

More information

Linear Classification. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Linear Classification. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington Linear Classification CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Example of Linear Classification Red points: patterns belonging

More information

Linear Models for Regression

Linear Models for Regression Linear Models for Regression CSE 4309 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 The Regression Problem Training data: A set of input-output

More information

Final Examination CS 540-2: Introduction to Artificial Intelligence

Final Examination CS 540-2: Introduction to Artificial Intelligence Final Examination CS 540-2: Introduction to Artificial Intelligence May 7, 2017 LAST NAME: SOLUTIONS FIRST NAME: Problem Score Max Score 1 14 2 10 3 6 4 10 5 11 6 9 7 8 9 10 8 12 12 8 Total 100 1 of 11

More information

MAE 493G, CpE 493M, Mobile Robotics. 6. Basic Probability

MAE 493G, CpE 493M, Mobile Robotics. 6. Basic Probability MAE 493G, CpE 493M, Mobile Robotics 6. Basic Probability Instructor: Yu Gu, Fall 2013 Uncertainties in Robotics Robot environments are inherently unpredictable; Sensors and data acquisition systems are

More information

Bayesian Classifiers and Probability Estimation. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington

Bayesian Classifiers and Probability Estimation. Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington Bayesian Classifiers and Probability Estimation Vassilis Athitsos CSE 4308/5360: Artificial Intelligence I University of Texas at Arlington 1 Data Space Suppose that we have a classification problem The

More information

CSE 473: Artificial Intelligence

CSE 473: Artificial Intelligence CSE 473: Artificial Intelligence Probability Steve Tanimoto University of Washington [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials

More information

Statistical Learning. Philipp Koehn. 10 November 2015

Statistical Learning. Philipp Koehn. 10 November 2015 Statistical Learning Philipp Koehn 10 November 2015 Outline 1 Learning agents Inductive learning Decision tree learning Measuring learning performance Bayesian learning Maximum a posteriori and maximum

More information

CSE 312 Final Review: Section AA

CSE 312 Final Review: Section AA CSE 312 TAs December 8, 2011 General Information General Information Comprehensive Midterm General Information Comprehensive Midterm Heavily weighted toward material after the midterm Pre-Midterm Material

More information

Constructive Artificial Intelligence

Constructive Artificial Intelligence Probabilistic Inference Daniel Polani School of Computer Science University of Hertfordshire November 24, 2013 All rights reserved. Permission is granted to copy and distribute these slides in full or

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

Logistics. Naïve Bayes & Expectation Maximization. 573 Schedule. Coming Soon. Estimation Models. Topics

Logistics. Naïve Bayes & Expectation Maximization. 573 Schedule. Coming Soon. Estimation Models. Topics Logistics Naïve Bayes & Expectation Maximization CSE 7 eam Meetings Midterm Open book, notes Studying See AIMA exercises Daniel S. Weld Daniel S. Weld 7 Schedule Selected opics Coming Soon Selected opics

More information

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes.

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes. CMSC 310 Artificial Intelligence Probabilistic Reasoning and Bayesian Belief Networks Probabilities, Random Variables, Probability Distribution, Conditional Probability, Joint Distributions, Bayes Theorem

More information

Introduction to Artificial Intelligence. Unit # 11

Introduction to Artificial Intelligence. Unit # 11 Introduction to Artificial Intelligence Unit # 11 1 Course Outline Overview of Artificial Intelligence State Space Representation Search Techniques Machine Learning Logic Probabilistic Reasoning/Bayesian

More information

Autonomous Mobile Robot Design

Autonomous Mobile Robot Design Autonomous Mobile Robot Design Topic: Particle Filter for Localization Dr. Kostas Alexis (CSE) These slides relied on the lectures from C. Stachniss, and the book Probabilistic Robotics from Thurn et al.

More information

Bayesian RL Seminar. Chris Mansley September 9, 2008

Bayesian RL Seminar. Chris Mansley September 9, 2008 Bayesian RL Seminar Chris Mansley September 9, 2008 Bayes Basic Probability One of the basic principles of probability theory, the chain rule, will allow us to derive most of the background material in

More information

Bayesian Updating: Discrete Priors: Spring 2014

Bayesian Updating: Discrete Priors: Spring 2014 ian Updating: Discrete Priors: 18.05 Spring 2014 http://xkcd.com/1236/ January 1, 2017 1 / 16 Learning from experience Which treatment would you choose? 1. Treatment 1: cured 100% of patients in a trial.

More information

Human-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg

Human-Oriented Robotics. Temporal Reasoning. Kai Arras Social Robotics Lab, University of Freiburg Temporal Reasoning Kai Arras, University of Freiburg 1 Temporal Reasoning Contents Introduction Temporal Reasoning Hidden Markov Models Linear Dynamical Systems (LDS) Kalman Filter 2 Temporal Reasoning

More information

Cryptography: Joining the RSA Cryptosystem

Cryptography: Joining the RSA Cryptosystem Cryptography: Joining the RSA Cryptosystem Greg Plaxton Theory in Programming Practice, Fall 2005 Department of Computer Science University of Texas at Austin Joining the RSA Cryptosystem: Overview First,

More information

Bayesian Networks. Machine Learning, Fall Slides based on material from the Russell and Norvig AI Book, Ch. 14

Bayesian Networks. Machine Learning, Fall Slides based on material from the Russell and Norvig AI Book, Ch. 14 Bayesian Networks Machine Learning, Fall 2010 Slides based on material from the Russell and Norvig AI Book, Ch. 14 1 Administrativia Bayesian networks The inference problem: given a BN, how to make predictions

More information

CSE 473: Artificial Intelligence Probability Review à Markov Models. Outline

CSE 473: Artificial Intelligence Probability Review à Markov Models. Outline CSE 473: Artificial Intelligence Probability Review à Markov Models Daniel Weld University of Washington [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.

More information

Probabilistic Reasoning. Kee-Eung Kim KAIST Computer Science

Probabilistic Reasoning. Kee-Eung Kim KAIST Computer Science Probabilistic Reasoning Kee-Eung Kim KAIST Computer Science Outline #1 Acting under uncertainty Probabilities Inference with Probabilities Independence and Bayes Rule Bayesian networks Inference in Bayesian

More information

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington Neural Networks CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Perceptrons x 0 = 1 x 1 x 2 z = h w T x Output: z x D A perceptron

More information

Learning Theory. Machine Learning CSE546 Carlos Guestrin University of Washington. November 25, Carlos Guestrin

Learning Theory. Machine Learning CSE546 Carlos Guestrin University of Washington. November 25, Carlos Guestrin Learning Theory Machine Learning CSE546 Carlos Guestrin University of Washington November 25, 2013 Carlos Guestrin 2005-2013 1 What now n We have explored many ways of learning from data n But How good

More information

CSEP 573: Artificial Intelligence

CSEP 573: Artificial Intelligence CSEP 573: Artificial Intelligence Bayesian Networks: Inference Ali Farhadi Many slides over the course adapted from either Luke Zettlemoyer, Pieter Abbeel, Dan Klein, Stuart Russell or Andrew Moore 1 Outline

More information

CS 343: Artificial Intelligence

CS 343: Artificial Intelligence CS 343: Artificial Intelligence Probability Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188

More information

Infinite Continued Fractions

Infinite Continued Fractions Infinite Continued Fractions 8-5-200 The value of an infinite continued fraction [a 0 ; a, a 2, ] is lim c k, where c k is the k-th convergent k If [a 0 ; a, a 2, ] is an infinite continued fraction with

More information

last two digits of your SID

last two digits of your SID Announcements Midterm: Wednesday 7pm-9pm See midterm prep page (posted on Piazza, inst.eecs page) Four rooms; your room determined by last two digits of your SID: 00-32: Dwinelle 155 33-45: Genetics and

More information

Text Categorization CSE 454. (Based on slides by Dan Weld, Tom Mitchell, and others)

Text Categorization CSE 454. (Based on slides by Dan Weld, Tom Mitchell, and others) Text Categorization CSE 454 (Based on slides by Dan Weld, Tom Mitchell, and others) 1 Given: Categorization A description of an instance, x X, where X is the instance language or instance space. A fixed

More information

Last Time. Today. Bayesian Learning. The Distributions We Love. CSE 446 Gaussian Naïve Bayes & Logistic Regression

Last Time. Today. Bayesian Learning. The Distributions We Love. CSE 446 Gaussian Naïve Bayes & Logistic Regression CSE 446 Gaussian Naïve Bayes & Logistic Regression Winter 22 Dan Weld Learning Gaussians Naïve Bayes Last Time Gaussians Naïve Bayes Logistic Regression Today Some slides from Carlos Guestrin, Luke Zettlemoyer

More information

Learning and Neural Networks

Learning and Neural Networks Artificial Intelligence Learning and Neural Networks Readings: Chapter 19 & 20.5 of Russell & Norvig Example: A Feed-forward Network w 13 I 1 H 3 w 35 w 14 O 5 I 2 w 23 w 24 H 4 w 45 a 5 = g 5 (W 3,5 a

More information

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)

More information

Priors, Total Probability, Expectation, Multiple Trials

Priors, Total Probability, Expectation, Multiple Trials Priors, Total Probability, Expectation, Multiple Trials CS 2800: Discrete Structures, Fall 2014 Sid Chaudhuri Bayes' Theorem Given: prior probabilities of hypotheses, and the probability that each hypothesis

More information

Introduction to Artificial Intelligence Belief networks

Introduction to Artificial Intelligence Belief networks Introduction to Artificial Intelligence Belief networks Chapter 15.1 2 Dieter Fox Based on AIMA Slides c S. Russell and P. Norvig, 1998 Chapter 15.1 2 0-0 Outline Bayesian networks: syntax and semantics

More information

Preliminary check: are the terms that we are adding up go to zero or not? If not, proceed! If the terms a n are going to zero, pick another test.

Preliminary check: are the terms that we are adding up go to zero or not? If not, proceed! If the terms a n are going to zero, pick another test. Throughout these templates, let series. be a series. We hope to determine the convergence of this Divergence Test: If lim is not zero or does not exist, then the series diverges. Preliminary check: are

More information

CSE 473: Artificial Intelligence Autumn Topics

CSE 473: Artificial Intelligence Autumn Topics CSE 473: Artificial Intelligence Autumn 2014 Bayesian Networks Learning II Dan Weld Slides adapted from Jack Breese, Dan Klein, Daphne Koller, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 473 Topics

More information

MATH 240. Chapter 8 Outlines of Hypothesis Tests

MATH 240. Chapter 8 Outlines of Hypothesis Tests MATH 4 Chapter 8 Outlines of Hypothesis Tests Test for Population Proportion p Specify the null and alternative hypotheses, ie, choose one of the three, where p is some specified number: () H : p H : p

More information

Grundlagen der Künstlichen Intelligenz

Grundlagen der Künstlichen Intelligenz Grundlagen der Künstlichen Intelligenz Uncertainty & Probabilities & Bandits Daniel Hennes 16.11.2017 (WS 2017/18) University Stuttgart - IPVS - Machine Learning & Robotics 1 Today Uncertainty Probability

More information

Artificial neural networks

Artificial neural networks Artificial neural networks Chapter 8, Section 7 Artificial Intelligence, spring 203, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 8, Section 7 Outline Brains Neural

More information

Adam Blank Spring 2017 CSE 311. Foundations of Computing I

Adam Blank Spring 2017 CSE 311. Foundations of Computing I Adam Blank Spring 2017 CSE 311 Foundations of Computing I Pre-Lecture Problem Suppose that p, and p (q r) are true. Is q true? Can you prove it with equivalences? CSE 311: Foundations of Computing Lecture

More information

COMP9414: Artificial Intelligence Reasoning Under Uncertainty

COMP9414: Artificial Intelligence Reasoning Under Uncertainty COMP9414, Monday 16 April, 2012 Reasoning Under Uncertainty 2 COMP9414: Artificial Intelligence Reasoning Under Uncertainty Overview Problems with Logical Approach What Do the Numbers Mean? Wayne Wobcke

More information

Probabilistic Robotics

Probabilistic Robotics University of Rome La Sapienza Master in Artificial Intelligence and Robotics Probabilistic Robotics Prof. Giorgio Grisetti Course web site: http://www.dis.uniroma1.it/~grisetti/teaching/probabilistic_ro

More information

Uncertainty. Chapter 13, Sections 1 6

Uncertainty. Chapter 13, Sections 1 6 Uncertainty Chapter 13, Sections 1 6 Artificial Intelligence, spring 2013, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 13, Sections 1 6 1 Outline Uncertainty Probability

More information

Our Status in CSE 5522

Our Status in CSE 5522 Our Status in CSE 5522 We re done with Part I Search and Planning! Part II: Probabilistic Reasoning Diagnosis Speech recognition Tracking objects Robot mapping Genetics Error correcting codes lots more!

More information

Learning Objectives. c D. Poole and A. Mackworth 2010 Artificial Intelligence, Lecture 7.2, Page 1

Learning Objectives. c D. Poole and A. Mackworth 2010 Artificial Intelligence, Lecture 7.2, Page 1 Learning Objectives At the end of the class you should be able to: identify a supervised learning problem characterize how the prediction is a function of the error measure avoid mixing the training and

More information

Probability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides

Probability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides Probability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides slide 1 Inference with Bayes rule: Example In a bag there are two envelopes one has a red ball (worth $100) and a black ball one

More information

Bayesian Methods: Naïve Bayes

Bayesian Methods: Naïve Bayes Bayesian Methods: aïve Bayes icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Last Time Parameter learning Learning the parameter of a simple coin flipping model Prior

More information

Introduction to Artificial Intelligence Midterm 2. CS 188 Spring You have approximately 2 hours and 50 minutes.

Introduction to Artificial Intelligence Midterm 2. CS 188 Spring You have approximately 2 hours and 50 minutes. CS 188 Spring 2014 Introduction to Artificial Intelligence Midterm 2 You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your two-page crib sheet. Mark your answers

More information

Today. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use?

Today. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use? Today Statistical Learning Parameter Estimation: Maximum Likelihood (ML) Maximum A Posteriori (MAP) Bayesian Continuous case Learning Parameters for a Bayesian Network Naive Bayes Maximum Likelihood estimates

More information

Discrete Structures - CM0246 Cardinality

Discrete Structures - CM0246 Cardinality Discrete Structures - CM0246 Cardinality Andrés Sicard-Ramírez Universidad EAFIT Semester 2014-2 Cardinality Definition (Cardinality (finite sets)) Let A be a set. The number of (distinct) elements in

More information

CS188 Outline. We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning. Part III: Machine Learning

CS188 Outline. We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning. Part III: Machine Learning CS188 Outline We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning Diagnosis Speech recognition Tracking objects Robot mapping Genetics Error correcting codes lots more! Part III:

More information

Lecture Slides. Elementary Statistics. by Mario F. Triola. and the Triola Statistics Series

Lecture Slides. Elementary Statistics. by Mario F. Triola. and the Triola Statistics Series Lecture Slides Elementary Statistics Tenth Edition and the Triola Statistics Series by Mario F. Triola Slide 1 Chapter 9 Inferences from Two Samples 9-1 Overview 9-2 Inferences About Two Proportions 9-3

More information

Hidden Markov Models Part 1: Introduction

Hidden Markov Models Part 1: Introduction Hidden Markov Models Part 1: Introduction CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Modeling Sequential Data Suppose that

More information

Outline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

Outline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car CSE 573: Artificial Intelligence Autumn 2012 Bayesian Networks Dan Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer Outline Probabilistic models (and inference)

More information

EQ: What are limits, and how do we find them? Finite limits as x ± Horizontal Asymptote. Example Horizontal Asymptote

EQ: What are limits, and how do we find them? Finite limits as x ± Horizontal Asymptote. Example Horizontal Asymptote Finite limits as x ± The symbol for infinity ( ) does not represent a real number. We use to describe the behavior of a function when the values in its domain or range outgrow all finite bounds. For example,

More information

Final Exam December 12, 2017

Final Exam December 12, 2017 Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes

More information

Biointelligence Lab School of Computer Sci. & Eng. Seoul National University

Biointelligence Lab School of Computer Sci. & Eng. Seoul National University Artificial Intelligence Chater 19 easoning with Uncertain Information Biointelligence Lab School of Comuter Sci. & Eng. Seoul National University Outline l eview of Probability Theory l Probabilistic Inference

More information

CSE 473: Artificial Intelligence Autumn 2011

CSE 473: Artificial Intelligence Autumn 2011 CSE 473: Artificial Intelligence Autumn 2011 Bayesian Networks Luke Zettlemoyer Many slides over the course adapted from either Dan Klein, Stuart Russell or Andrew Moore 1 Outline Probabilistic models

More information

CS 343: Artificial Intelligence

CS 343: Artificial Intelligence CS 343: Artificial Intelligence Bayes Nets: Sampling Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.

More information

Computational Cognitive Science

Computational Cognitive Science Computational Cognitive Science Lecture 8: Frank Keller School of Informatics University of Edinburgh keller@inf.ed.ac.uk Based on slides by Sharon Goldwater October 14, 2016 Frank Keller Computational

More information

Probability. CS 3793/5233 Artificial Intelligence Probability 1

Probability. CS 3793/5233 Artificial Intelligence Probability 1 CS 3793/5233 Artificial Intelligence 1 Motivation Motivation Random Variables Semantics Dice Example Joint Dist. Ex. Axioms Agents don t have complete knowledge about the world. Agents need to make decisions

More information

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Some slides are due to Christopher Bishop Limitations of K-means Hard assignments of data points to clusters small shift of a

More information

MODULE 10 Bayes Classifier LESSON 20

MODULE 10 Bayes Classifier LESSON 20 MODULE 10 Bayes Classifier LESSON 20 Bayesian Belief Networks Keywords: Belief, Directed Acyclic Graph, Graphical Model 1 Bayesian Belief Network A Bayesian network is a graphical model of a situation

More information

Model Averaging (Bayesian Learning)

Model Averaging (Bayesian Learning) Model Averaging (Bayesian Learning) We want to predict the output Y of a new case that has input X = x given the training examples e: p(y x e) = m M P(Y m x e) = m M P(Y m x e)p(m x e) = m M P(Y m x)p(m

More information

INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. NPTEL National Programme on Technology Enhanced Learning. Probability Methods in Civil Engineering

INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR. NPTEL National Programme on Technology Enhanced Learning. Probability Methods in Civil Engineering INDIAN INSTITUTE OF TECHNOLOGY KHARAGPUR NPTEL National Programme on Technology Enhanced Learning Probability Methods in Civil Engineering Prof. Rajib Maity Department of Civil Engineering IIT Kharagpur

More information

Bayes Theorem (10B) Young Won Lim 6/3/17

Bayes Theorem (10B) Young Won Lim 6/3/17 Bayes Theorem (10B) Copyright (c) 2017 Young W. Lim. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any later

More information

Sec$on Summary. Mathematical Proofs Forms of Theorems Trivial & Vacuous Proofs Direct Proofs Indirect Proofs

Sec$on Summary. Mathematical Proofs Forms of Theorems Trivial & Vacuous Proofs Direct Proofs Indirect Proofs Section 1.7 Sec$on Summary Mathematical Proofs Forms of Theorems Trivial & Vacuous Proofs Direct Proofs Indirect Proofs Proof of the Contrapositive Proof by Contradiction 2 Proofs of Mathema$cal Statements

More information

Bayes Theorem (4A) Young Won Lim 3/5/18

Bayes Theorem (4A) Young Won Lim 3/5/18 Bayes Theorem (4A) Copyright (c) 2017-2018 Young W. Lim. Permission is granted to copy, distribute and/or modify this document under the terms of the GNU Free Documentation License, Version 1.2 or any

More information

CS 231A Section 1: Linear Algebra & Probability Review

CS 231A Section 1: Linear Algebra & Probability Review CS 231A Section 1: Linear Algebra & Probability Review 1 Topics Support Vector Machines Boosting Viola-Jones face detector Linear Algebra Review Notation Operations & Properties Matrix Calculus Probability

More information

Final Exam December 12, 2017

Final Exam December 12, 2017 Introduction to Artificial Intelligence CSE 473, Autumn 2017 Dieter Fox Final Exam December 12, 2017 Directions This exam has 7 problems with 111 points shown in the table below, and you have 110 minutes

More information

CS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang

CS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang CS 231A Section 1: Linear Algebra & Probability Review Kevin Tang Kevin Tang Section 1-1 9/30/2011 Topics Support Vector Machines Boosting Viola Jones face detector Linear Algebra Review Notation Operations

More information

Computational Perception. Bayesian Inference

Computational Perception. Bayesian Inference Computational Perception 15-485/785 January 24, 2008 Bayesian Inference The process of probabilistic inference 1. define model of problem 2. derive posterior distributions and estimators 3. estimate parameters

More information

CS 188: Artificial Intelligence. Our Status in CS188

CS 188: Artificial Intelligence. Our Status in CS188 CS 188: Artificial Intelligence Probability Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. 1 Our Status in CS188 We re done with Part I Search and Planning! Part II: Probabilistic Reasoning

More information

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein Kalman filtering and friends: Inference in time series models Herke van Hoof slides mostly by Michael Rubinstein Problem overview Goal Estimate most probable state at time k using measurement up to time

More information

Be able to define the following terms and answer basic questions about them:

Be able to define the following terms and answer basic questions about them: CS440/ECE448 Section Q Fall 2017 Final Review Be able to define the following terms and answer basic questions about them: Probability o Random variables, axioms of probability o Joint, marginal, conditional

More information

Algorithm Analysis Recurrence Relation. Chung-Ang University, Jaesung Lee

Algorithm Analysis Recurrence Relation. Chung-Ang University, Jaesung Lee Algorithm Analysis Recurrence Relation Chung-Ang University, Jaesung Lee Recursion 2 Recursion 3 Recursion in Real-world Fibonacci sequence = + Initial conditions: = 0 and = 1. = + = + = + 0, 1, 1, 2,

More information

Probability Basics. Robot Image Credit: Viktoriya Sukhanova 123RF.com

Probability Basics. Robot Image Credit: Viktoriya Sukhanova 123RF.com Probability Basics These slides were assembled by Eric Eaton, with grateful acknowledgement of the many others who made their course materials freely available online. Feel free to reuse or adapt these

More information

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation CS 188: Artificial Intelligence Spring 2010 Lecture 15: Bayes Nets II Independence 3/9/2010 Pieter Abbeel UC Berkeley Many slides over the course adapted from Dan Klein, Stuart Russell, Andrew Moore Current

More information

Bayes Nets III: Inference

Bayes Nets III: Inference 1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy

More information

Bayesian Learning. Chapter 6: Bayesian Learning. Bayes Theorem. Roles for Bayesian Methods. CS 536: Machine Learning Littman (Wu, TA)

Bayesian Learning. Chapter 6: Bayesian Learning. Bayes Theorem. Roles for Bayesian Methods. CS 536: Machine Learning Littman (Wu, TA) Bayesian Learning Chapter 6: Bayesian Learning CS 536: Machine Learning Littan (Wu, TA) [Read Ch. 6, except 6.3] [Suggested exercises: 6.1, 6.2, 6.6] Bayes Theore MAP, ML hypotheses MAP learners Miniu

More information

Introduction to Spring 2006 Artificial Intelligence Practice Final

Introduction to Spring 2006 Artificial Intelligence Practice Final NAME: SID#: Login: Sec: 1 CS 188 Introduction to Spring 2006 Artificial Intelligence Practice Final You have 180 minutes. The exam is open-book, open-notes, no electronics other than basic calculators.

More information

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22. Introduction to Artificial Intelligence V22.0472-001 Fall 2009 Lecture 15: Bayes Nets 3 Midterms graded Assignment 2 graded Announcements Rob Fergus Dept of Computer Science, Courant Institute, NYU Slides

More information

Mathematical Induction. Defining Functions. Overview. Notation for recursive functions. Base case Sn(0) = 0 S(n) = S(n 1) + n for n > 0

Mathematical Induction. Defining Functions. Overview. Notation for recursive functions. Base case Sn(0) = 0 S(n) = S(n 1) + n for n > 0 Readings on induction. Mathematical Induction (a) Weiss, Sec. 7.2, page 233 (b) Course slides f lecture and notes recitation. Every criticism from a good man is of value to me. What you hint at generally

More information

Continuous Data with Continuous Priors Class 14, Jeremy Orloff and Jonathan Bloom

Continuous Data with Continuous Priors Class 14, Jeremy Orloff and Jonathan Bloom 1 Learning Goals Continuous Data with Continuous Priors Class 14, 18.0 Jeremy Orloff and Jonathan Bloom 1. Be able to construct a ian update table for continuous hypotheses and continuous data. 2. Be able

More information

CSE 21 Practice Exam for Midterm 2 Fall 2017

CSE 21 Practice Exam for Midterm 2 Fall 2017 CSE 1 Practice Exam for Midterm Fall 017 These practice problems should help prepare you for the second midterm, which is on monday, November 11 This is longer than the actual exam will be, but good practice

More information

Sample solutions to Homework 4, Information-Theoretic Modeling (Fall 2014)

Sample solutions to Homework 4, Information-Theoretic Modeling (Fall 2014) Sample solutions to Homework 4, Information-Theoretic Modeling (Fall 204) Jussi Määttä October 2, 204 Question [First, note that we use the symbol! as an end-of-message symbol. When we see it, we know

More information