CS 331: Artificial Intelligence Probability I. Dealing with Uncertainty

Similar documents
Dealing with Uncertainty. CS 331: Artificial Intelligence Probability I. Outline. Random Variables. Random Variables.

Probability Basics. Robot Image Credit: Viktoriya Sukhanova 123RF.com

Probabilistic and Bayesian Analytics

Sources of Uncertainty

CS 4649/7649 Robot Intelligence: Planning

Basic Probability and Statistics

Machine Learning

CISC 4631 Data Mining

Machine Learning

Probability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides

Machine Learning

Bayesian Reasoning. Adapted from slides by Tim Finin and Marie desjardins.

Why Probability? It's the right way to look at the world.

Artificial Intelligence

Uncertainty. Chapter 13

Probability Theory Review

Bayes Nets for representing

Uncertainty. Chapter 13

Basic Probabilistic Reasoning SEG

Uncertainty and knowledge. Uncertainty and knowledge. Reasoning with uncertainty. Notes

Uncertainty. Outline

ARTIFICIAL INTELLIGENCE. Uncertainty: probabilistic reasoning

CS 561: Artificial Intelligence

CS 5100: Founda.ons of Ar.ficial Intelligence

CS 188: Artificial Intelligence Spring Announcements

Discrete Random Variables

Refresher on Discrete Probability

Bayes Nets for representing and reasoning about uncertainty

Aarti Singh. Lecture 2, January 13, Reading: Bishop: Chap 1,2. Slides courtesy: Eric Xing, Andrew Moore, Tom Mitchell

STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas. Propositional Logic and Probability Theory: Review

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes.

COMP9414/ 9814/ 3411: Artificial Intelligence. 14. Uncertainty. Russell & Norvig, Chapter 13. UNSW c AIMA, 2004, Alan Blair, 2012

CS 188: Artificial Intelligence. Bayes Nets

Axioms of Probability? Notation. Bayesian Networks. Bayesian Networks. Today we ll introduce Bayesian Networks.

An AI-ish view of Probability, Conditional Probability & Bayes Theorem

10/18/2017. An AI-ish view of Probability, Conditional Probability & Bayes Theorem. Making decisions under uncertainty.

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

Probability and Uncertainty. Bayesian Networks

CS188 Outline. We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning. Part III: Machine Learning

CS 343: Artificial Intelligence

Lecture 8. Probabilistic Reasoning CS 486/686 May 25, 2006

Uncertainty. Introduction to Artificial Intelligence CS 151 Lecture 2 April 1, CS151, Spring 2004

A Problem Involving Games. Paccioli s Solution. Problems for Paccioli: Small Samples. n / (n + m) m / (n + m)

Basic Probability. Robert Platt Northeastern University. Some images and slides are used from: 1. AIMA 2. Chris Amato

Our Status in CSE 5522

Resolution or modus ponens are exact there is no possibility of mistake if the rules are followed exactly.

Introduction to Probabilistic Reasoning. Image credit: NASA. Assignment

Vibhav Gogate University of Texas, Dallas

UNCERTAINTY. In which we see what an agent should do when not all is crystal-clear.

Fundamentals of Probability CE 311S

Where are we in CS 440?

Probabilistic representation and reasoning

Review of Basic Probability

Random Variables and Events

Announcements. CS 188: Artificial Intelligence Spring Bayes Net Semantics. Probabilities in BNs. All Conditional Independences

CMPSCI 240: Reasoning about Uncertainty

Chapter 13 Quantifying Uncertainty

Basic notions of probability theory

Stochastic Methods. 5.0 Introduction 5.1 The Elements of Counting 5.2 Elements of Probability Theory

Symbolic Logic Outline

Where are we? Knowledge Engineering Semester 2, Reasoning under Uncertainty. Probabilistic Reasoning

COMP5211 Lecture Note on Reasoning under Uncertainty

Probabilistic Robotics

Probabilistic Reasoning

Fusion in simple models

Probabilistic and Bayesian Analytics Based on a Tutorial by Andrew W. Moore, Carnegie Mellon University

Reasoning under Uncertainty: Intro to Probability

Where are we? è Five major sec3ons of this course

Probabilistic representation and reasoning

Probability. CS 3793/5233 Artificial Intelligence Probability 1

Basic notions of probability theory

Artificial Intelligence Uncertainty

Information Science 2

Objectives. Probabilistic Reasoning Systems. Outline. Independence. Conditional independence. Conditional independence II.

Probability theory: elements

Uncertainty in the World. Representing Uncertainty. Uncertainty in the World and our Models. Uncertainty

Uncertainty (Chapter 13, Russell & Norvig) Introduction to Artificial Intelligence CS 150 Lecture 14

UVA CS 6316/4501 Fall 2016 Machine Learning. Lecture 11: Probability Review. Dr. Yanjun Qi. University of Virginia. Department of Computer Science

Probabilistic Reasoning Systems

22c:145 Artificial Intelligence

Where are we in CS 440?

Lecture 1: Probability Fundamentals

Outline. Uncertainty. Methods for handling uncertainty. Uncertainty. Making decisions under uncertainty. Probability. Uncertainty

CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-II

Probabilistic Representation and Reasoning

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

CS 5522: Artificial Intelligence II

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Bayes and Naïve Bayes Classifiers CS434

Uncertainty. Outline. Probability Syntax and Semantics Inference Independence and Bayes Rule. AIMA2e Chapter 13

Reinforcement Learning Wrap-up

Bayes Nets for representing and reasoning about uncertainty

CSEP 573: Artificial Intelligence

Origins of Probability Theory

Reasoning under uncertainty

Uncertainty. Chapter 13, Sections 1 6

Web-Mining Agents Data Mining

Independence. CS 109 Lecture 5 April 6th, 2016

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation

Transcription:

CS 331: Artificial Intelligence Probability I Thanks to Andrew Moore for some course material 1 Dealing with Uncertainty We want to get to the point where we can reason with uncertainty This will require using probability eg. probability that it will rain today is 0.99 We will review the fundamentals of probability 2 1

1. Random variables 2. Probability Outline Random Variables The basic element of probability is the random variable Think of the random variable as an event with some degree of uncertainty as to whether that event occurs Random variables have a domain of values it can take on 4 2

Random Variables 3 types of random variables: 1. Boolean random variables 2. Discrete random variables 3. Continuous random variables Random Variables Example: ProfLate is a random variable for whether your prof will be late to class or not The domain of ProfLate is {true, false} ProfLate = true: proposition that prof will be late to class ProfLate = false: proposition that prof will not be late to class 6 3

Random Variables Example: ProfLate is a random variable for whether your prof will be late to class or not The domain of ProfLate is <true, false> ProfLate = true: proposition that prof will be late to class You can assign some degree of ProfLate = false: belief proposition to this proposition that eg. prof will not be late to class P(ProfLate = true) = 0.9 7 Random Variables Example: ProfLate is a random variable for whether your prof will be late to class or not The domain of ProfLate is <true, false> ProfLate = true: proposition that prof will be late to class ProfLate = false: proposition that prof will not be late to class And to this one eg. P(ProfLate = false) = 0.1 8 4

Random Variables We will refer to random variables with capitalized names eg. X, Y, ProfLate We will refer to names of values with lower case names eg. x, y, proflate This means you may see a statement like ProfLate = proflate This means the random variable ProfLate takes the value proflate (which can be true or false) Shorthand notation: ProfLate = true is the same as proflate and ProfLate = false is the same as proflate 9 Boolean Random Variables Take the values true or false Eg. Let A be a Boolean random variable P(A = false) = 0.9 P(A = true) = 0.1 10 5

Discrete Random Variables Allowed to taken on a finite number of values eg. P(DrinkSize=Small) = 0.1 P(DrinkSize=Medium) = 0.2 P(DrinkSize=Large) = 0.7 Discrete Random Variables Values of the domain must be: Mutually Exclusive ie. P( A = v i AND A = v j ) = 0 if i j This means, for instance, that you can t have a drink that is both Small and Medium Exhaustive ie. P(A = v 1 OR A = v 2 OR... OR A = v k ) = 1 This means that a drink can only be either small, medium or large. There isn t an extra large 6

Discrete Random Variables Values of the domain must be: Mutually Exclusive ie. P( A = v i AND A = v j ) = 0 if i j This means, for instance, The AND that here you means can t intersection have a drink that is both Smallie. and (A Medium = v i ) (A = v j ) Exhaustive ie. P(A = v 1 OR A = v 2 OR... OR A = v k ) = 1 This means that a drink can only be either small, The OR here means union ie. (A = v medium or large. There isn t an extra large 1 ) (A = v 2 )... (A = v k ) Discrete Random Variables Since we now have multi-valued discrete random variables we can t write P(a) or P( a) anymore We have to write P(A = v i ) where v i = a value in {v 1, v 2,, v k } 14 7

Continuous Random Variables Can take values from the real numbers eg. They can take values from [0, 1] Note: We will primarily be dealing with discrete random variables (The next slide is just to provide a little bit of information about continuous random variables) 15 Probability Density Functions Discrete random variables have probability distributions: 1.0 P(A) a a Continuous random variables have probability density functions eg: P(X) P(X) X X 8

Probabilities We will write P(A=true) as the fraction of possible worlds in which A is true We can debate the philosophical implications of this for the next 4 hours But we won t Probabilities We will sometimes talk about the probabilities of all possible values of a random variable Instead of writing P(A=false) = 0.25 P(A=true) = 0.75 We will write P(A) = (0.25, 0.75) Note the boldface! 18 9

Visualizing A Event space of all possible worlds Worlds in which A is true P(a) = Area of reddish oval Its area is 1 Worlds in which A is false 19 The Axioms of Probability 0 P(a) 1 P(true) = 1 P(false) = 0 P(aOR b) = P(a) + P(b) - P(a AND b) The logical OR is equivalent to set union. The logical AND is equivalent to set intersection ( ). Sometimes, I ll write it as P(a, b) These axioms are often called Kolmogorov s axioms in honor of the Russian mathematician Andrei Kolmogorov 20 10

Interpreting the axioms 0 P(a) = 1 P(true) = 1 P(false) = 0 P(a OR b) = P(a) + P(b) - P(a, b) The area of P(a) can t get any smaller than 0 And a zero area would mean that there is no world in which a is false 21 Interpreting the axioms 0 P(a) 1 P(true) = 1 P(false) = 0 P(a OR b) = P(a) + P(b) - P(a, b) The area of P(a) can t get any bigger than 1 And an area of 1 would mean all worlds will have a is true 22 11

Interpreting the axioms 0 P(a) 1 P(true) = 1 P(false) = 0 P(a OR b) = P(a) + P(b) - P(a, b) P(a, b) [The purple area] a b P(a OR b) [the area of both circles) 23 These Axioms are Not to be Trifled With There have been attempts to do different methodologies for uncertainty Fuzzy Logic Three-valued logic Dempster-Shafer Non-monotonic reasoning But the axioms of probability are the only system with this property: If you gamble using them you can t be unfairly exploited by an opponent using some other system [de Finetti 1931] 12

Prior Probability We can consider P(A) as the unconditional or prior probability eg. P(ProfLate = true) = 1.0 It is the probability of event A in the absence of any other information If we get new information that affects A, we can reason with the conditional probability of A given the new information. 25 Conditional Probability P(A B) = Fraction of worlds in which B is true that also have A true Read this as: Probability of A conditioned on B Prior probability P(A) is a special case of the conditional probability P(A ) conditioned on no evidence 26 13

Conditional Probability Example F H = Have a headache F = Coming down with Flu H P(H) = 1/10 P(F) = 1/40 P(H F) = 1/2 Headaches are rare and flu is rarer, but if you re coming down with flu there s a 50-50 chance you ll have a headache. 27 Conditional Probability F H = Have a headache F = Coming down with Flu P(H) = 1/10 P(F) = 1/40 P(H F) = 1/2 H P(H F) = Fraction of flu-inflicted worlds in which you have a headache # worlds with flu and headache = # worlds with flu Area of "H and F" region = Area of "F" region P(H, F) = P(F) 28 14

Definition of Conditional Probability P ( A B) = P( A, B) P( B) Corollary: The Chain Rule (aka The Product Rule) P ( A, B) = P( A B) P( B) 29 Important Note P( A B) + P( A B) = 1 But: P( A B) + P( A B) does not always = 1 30 15

The Joint Probability Distribution P(A, B ) is called the joint probability distribution of A and B It captures the probabilities of all combinations of the values of a set of random variables 31 The Joint Probability Distribution For example, if A and B are Boolean random variables, then P(A,B) could be specified as: P(A=false, B=false) 0.25 P(A=false, B=true) 0.25 P(A=true, B=false) 0.25 P(A=true, B=true) 0.25 32 16

The Joint Probability Distribution Now suppose we have the random variables: Drink = {Coke, Sprite} Size = {Small, Medium, Large} The joint probability distribution for P(Drink,Size) could look like: P(Drink=Coke, Size=Small) 0.1 P(Drink=Coke, Size=Medium) 0.1 P(Drink=Coke, Size=Large) 0.3 P(Drink=Sprite, Size=Small) 0.1 P(Drink=Sprite, Size=Medium) 0.2 P(Drink=Sprite, Size=Large) 0.2 33 Full Joint Probability Distribution Suppose you have the complete set of random variables used to describe the world A joint probability distribution that covers this complete set is called the full joint probability distribution Is a complete specification of one s uncertainty about the world in question Very powerful: Can be used to answer any probabilistic query 34 17