Where are we in CS 440?

Similar documents
Where are we in CS 440?

n How to represent uncertainty in knowledge? n Which action to choose under uncertainty? q Assume the car does not have a flat tire

Uncertainty. Outline

Artificial Intelligence

Uncertainty. Russell & Norvig Chapter 13.

Uncertainty. Introduction to Artificial Intelligence CS 151 Lecture 2 April 1, CS151, Spring 2004

Uncertainty. Chapter 13

Artificial Intelligence Uncertainty

Uncertainty. Chapter 13

Pengju XJTU 2016

Probabilistic Robotics

Probabilistic Reasoning

An AI-ish view of Probability, Conditional Probability & Bayes Theorem

10/18/2017. An AI-ish view of Probability, Conditional Probability & Bayes Theorem. Making decisions under uncertainty.

Uncertain Knowledge and Reasoning

Uncertainty. Outline. Probability Syntax and Semantics Inference Independence and Bayes Rule. AIMA2e Chapter 13

Outline. Uncertainty. Methods for handling uncertainty. Uncertainty. Making decisions under uncertainty. Probability. Uncertainty

CS 561: Artificial Intelligence

COMP9414/ 9814/ 3411: Artificial Intelligence. 14. Uncertainty. Russell & Norvig, Chapter 13. UNSW c AIMA, 2004, Alan Blair, 2012

Uncertainty and Bayesian Networks

Uncertainty. Chapter 13. Chapter 13 1

Probabilistic Reasoning

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

Web-Mining Agents Data Mining

Basic Probability and Decisions

Uncertainty (Chapter 13, Russell & Norvig) Introduction to Artificial Intelligence CS 150 Lecture 14

UNCERTAINTY. In which we see what an agent should do when not all is crystal-clear.

Uncertainty. Chapter 13, Sections 1 6

Quantifying uncertainty & Bayesian networks

Uncertainty. Logic and Uncertainty. Russell & Norvig. Readings: Chapter 13. One problem with logical-agent approaches: C:145 Artificial

Pengju

Lecture Overview. Introduction to Artificial Intelligence COMP 3501 / COMP Lecture 11: Uncertainty. Uncertainty.

Quantifying Uncertainty & Probabilistic Reasoning. Abdulla AlKhenji Khaled AlEmadi Mohammed AlAnsari

Artificial Intelligence CS 6364

Uncertainty (Chapter 13, Russell & Norvig)

Ch.6 Uncertain Knowledge. Logic and Uncertainty. Representation. One problem with logical approaches: Department of Computer Science

Chapter 13 Quantifying Uncertainty

Uncertainty. 22c:145 Artificial Intelligence. Problem of Logic Agents. Foundations of Probability. Axioms of Probability

Answering. Question Answering: Result. Application of the Theorem Prover: Question. Overview. Question Answering: Example.

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes.

Resolution or modus ponens are exact there is no possibility of mistake if the rules are followed exactly.

Uncertainty. Variables. assigns to each sentence numerical degree of belief between 0 and 1. uncertainty

Artificial Intelligence Programming Probability

Computer Science CPSC 322. Lecture 18 Marginalization, Conditioning

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Our Status. We re done with Part I Search and Planning!

1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ).

CS 188: Artificial Intelligence. Our Status in CS188

CS188 Outline. CS 188: Artificial Intelligence. Today. Inference in Ghostbusters. Probability. We re done with Part I: Search and Planning!

Probabilistic Robotics. Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics (S. Thurn et al.

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10

Probabilistic representation and reasoning

CS 188: Artificial Intelligence Fall 2009

Probability Hal Daumé III. Computer Science University of Maryland CS 421: Introduction to Artificial Intelligence 27 Mar 2012

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018

ARTIFICIAL INTELLIGENCE. Uncertainty: probabilistic reasoning

Reasoning with Uncertainty. Chapter 13

Probability and Decision Theory

: A sea change in AI technologies

PROBABILITY AND INFERENCE

Discrete Probability and State Estimation

Discrete Probability and State Estimation

Random Variables. A random variable is some aspect of the world about which we (may) have uncertainty

Probabilistic Reasoning. Kee-Eung Kim KAIST Computer Science

Course Introduction. Probabilistic Modelling and Reasoning. Relationships between courses. Dealing with Uncertainty. Chris Williams.

CS 361: Probability & Statistics

Lecture 2. Conditional Probability

Objectives. Probabilistic Reasoning Systems. Outline. Independence. Conditional independence. Conditional independence II.

Probability theory: elements

Reasoning under Uncertainty: Intro to Probability

Probabilistic Reasoning Systems

Introduction to Probability Theory, Algebra, and Set Theory

CS 188: Artificial Intelligence Fall 2011

Modeling and reasoning with uncertainty

Machine Learning. CS Spring 2015 a Bayesian Learning (I) Uncertainty

Cartesian-product sample spaces and independence

Probability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides

Lecture 1: Basics of Probability

CSE 473: Artificial Intelligence

In today s lecture. Conditional probability and independence. COSC343: Artificial Intelligence. Curse of dimensionality.

EE126: Probability and Random Processes

Probabilistic Models

Probability theory basics

CS 188: Artificial Intelligence Spring Announcements

Elementary Discrete Probability

P (A B) P ((B C) A) P (B A) = P (B A) + P (C A) P (A) = P (B A) + P (C A) = Q(A) + Q(B).

2.4 Conditional Probability

Great Theoretical Ideas in Computer Science

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Reasoning under Uncertainty: Intro to Probability

Uncertainty and Belief Networks. Introduction to Artificial Intelligence CS 151 Lecture 1 continued Ok, Lecture 2!

CS 5522: Artificial Intelligence II

Y. Xiang, Inference with Uncertain Knowledge 1

Lecture 3 Probability Basics

Basic Probability and Statistics

Lecture 1: Review of Probability

CS 188: Artificial Intelligence Spring Announcements

An Introduction to Bayesian Reasoning

PROBABILISTIC REASONING SYSTEMS

Conditional probability

Transcription:

Where are we in CS 440? Now leaving: sequential deterministic reasoning Entering: probabilistic reasoning and machine learning

robability: Review of main concepts Chapter 3

Motivation: lanning under uncertainty Recall: representation for planning States are specified as conjunctions of predicates Start state: t CMI lane irportcmi irportord Goal state: t ORD ctions are described in terms of preconditions and effects: Flyp source dest recond: tp source lanep irportsource irportdest Effect: tp source tp dest

Motivation: lanning under uncertainty Let action t leave for airport t minutes before flight Will t succeed i.e. get me to the airport in time for the flight? roblems: artial observability road state other drivers' plans etc. Noisy sensors traffic reports Uncertainty in action outcomes flat tire etc. Complexity of modeling and predicting traffic Hence a purely logical approach either Risks falsehood: 25 will get me there on time or Leads to conclusions that are too weak for decision making: 25 will get me there on time if there's no accident on the bridge and it doesn't rain and my tires remain intact etc. etc. 440 will get me there on time but I ll have to stay overnight in the airport

robability robabilistic assertions summarize effects of Laziness: reluctance to enumerate exceptions qualifications etc. Ignorance: lack of explicit theories relevant facts initial conditions etc. Intrinsically random phenomena

Making decisions under uncertainty Suppose the agent believes the following: 25 gets me there on time 0.04 90 gets me there on time 0.70 20 gets me there on time 0.95 440 gets me there on time 0.9999 Which action should the agent choose? Depends on preferences for missing flight vs. time spent waiting Encapsulated by a utility function The agent should choose the action that maximizes the expected utility: t succeeds * U t succeeds + t fails * U t fails

Making decisions under uncertainty More generally: the expected utility of an action is defined as: EUa Σ outcomes of a outcome a Uoutcome Utility theory is used to represent and infer preferences Decision theory probability theory + utility theory

Monty Hall problem You re a contestant on a game show. You see three closed doors and behind one of them is a prize. You choose one door and the host opens one of the other doors and reveals that there is no prize behind it. Then he offers you a chance to switch to the remaining door. Should you take it? http://en.wikipedia.org/wiki/monty_hall_problem

Monty Hall problem With probability /3 you picked the correct door and with probability 2/3 picked the wrong door. If you picked the correct door and then you switch you lose. If you picked the wrong door and then you switch you win the prize. Expected utility of switching: EUSwitch /3 * 0 + 2/3 * rize Expected utility of not switching: EUNot switch /3 * rize + 2/3 * 0

Where do probabilities come from? Frequentism robabilities are relative frequencies For example if we toss a coin many times heads is the proportion of the time the coin will come up heads ut what if we re dealing with events that only happen once? E.g. what is the probability that Team X will win the Superbowl this year? Reference class problem Subjectivism robabilities are degrees of belief ut then how do we assign belief values to statements? What would constrain agents to hold consistent beliefs?

robabilities and rationality Why should a rational agent hold beliefs that are consistent with axioms of probability? For example + If an agent has some degree of belief in proposition he/she should be able to decide whether or not to accept a bet for/against De Finetti 93: If the agent believes that 0.4 should he/she agree to bet $4 that will occur against $6 that will not occur? Theorem: n agent who holds beliefs inconsistent with axioms of probability can be convinced to accept a combination of bets that is guaranteed to lose them money

Random variables We describe the uncertain state of the world using random variables Denoted by capital letters R: Is it raining? W: What s the weather? D: What is the outcome of rolling two dice? S: What is the speed of my car in MH? Just like variables in CSs random variables take on values in a domain Domain values must be mutually exclusive and exhaustive R in {True False} W in {Sunny Cloudy Rainy Snow} D in { 2 66} S in [0 200]

Events robabilistic statements are defined over events or sets of world states It is raining The weather is either cloudy or snowy The sum of the two dice rolls is My car is going between 30 and 50 miles per hour Events are described using propositions about random variables: R True W Cloudy W Snowy D {56 65} 30 S 50 Notation: is the probability of the set of world states in which proposition holds

Kolmogorov s axioms of probability For any propositions events 0 True and False 0 + Subtraction accounts for double-counting ased on these axioms what is? These axioms are sufficient to completely specify probability theory for discrete random variables For continuous variables need density functions

tomic events tomic event: a complete specification of the state of the world or a complete assignment of domain values to all random variables tomic events are mutually exclusive and exhaustive E.g. if the world consists of only two oolean variables Cavity and Toothache then there are four distinct atomic events: Cavity false Toothache false Cavity false Toothache true Cavity true Toothache false Cavity true Toothache true

Joint probability distributions joint distribution is an assignment of probabilities to every possible atomic event tomic event Cavity false Toothache false 0.8 Cavity false Toothache true 0. Cavity true Toothache false 0.05 Cavity true Toothache true 0.05 Why does it follow from the axioms of probability that the probabilities of all possible atomic events must sum to?

Joint probability distributions joint distribution is an assignment of probabilities to every possible atomic event Suppose we have a joint distribution of n random variables with domain sizes d What is the size of the probability table? Impossible to write out completely for all but the smallest distributions

Notation X x X 2 x 2 X n x n refers to a single entry atomic event in the joint probability distribution table Shorthand: x x 2 x n X X 2 X n refers to the entire joint probability distribution table can also refer to the probability of an event E.g. X x is an event

Marginal probability distributions From the joint distribution XY we can find the marginal distributions X and Y Cavity Toothache Cavity false Toothache false 0.8 Cavity false Toothache true 0. Cavity true Toothache false 0.05 Cavity true Toothache true 0.05 Cavity Cavity false? Cavity true? Toothache Toothache false? Toochache true?

Marginal probability distributions From the joint distribution XY we can find the marginal distributions X and Y To find X x sum the probabilities of all atomic events where X x: This is called marginalization we are marginalizing out all the variables except X n i i n n y x y x y x y Y x X y Y x X x X

Conditional probability robability of cavity given toothache: Cavity true Toothache true For any two events and

Conditional probability Cavity Toothache Cavity false Toothache false 0.8 Cavity false Toothache true 0. Cavity true Toothache false 0.05 Cavity true Toothache true 0.05 Cavity Cavity false 0.9 Cavity true 0. Toothache Toothache false 0.85 Toothache true 0.5 What is Cavity true Toothache false? 0.05 / 0.85 0.059 What is Cavity false Toothache true? 0. / 0.5 0.667

Conditional distributions conditional distribution is a distribution over the values of one variable given fixed values of other variables Cavity Toothache Cavity false Toothache false 0.8 Cavity false Toothache true 0. Cavity true Toothache false 0.05 Cavity true Toothache true 0.05 Cavity Toothache true Cavity false 0.667 Cavity true 0.333 Toothache Cavity true Toothache false 0.5 Toothache true 0.5 Cavity Toothache false Cavity false 0.94 Cavity true 0.059 Toothache Cavity false Toothache false 0.889 Toothache true 0.

Normalization trick To get the whole conditional distribution X Y y at once select all entries in the joint distribution table matching Y y and renormalize them to sum to one Cavity Toothache Cavity false Toothache false 0.8 Cavity false Toothache true 0. Cavity true Toothache false 0.05 Cavity true Toothache true 0.05 Toothache Cavity false Toothache false 0.8 Toothache true 0. Toothache Cavity false Select Renormalize Toothache false 0.889 Toothache true 0.

Normalization trick To get the whole conditional distribution X Y y at once select all entries in the joint distribution table matching Y y and renormalize them to sum to one Why does it work? y y x y x y x x ʹ ʹ by marginalization

roduct rule Definition of conditional probability: Sometimes we have the conditional probability and want to obtain the joint:

roduct rule Definition of conditional probability: Sometimes we have the conditional probability and want to obtain the joint: The chain rule: n i i i n n n 2 3 2

The irthday problem We have a set of n people. What is the probability that two of them share the same birthday? Easier to calculate the probability that n people do not share the same birthday distinct distinct distinct from distinct n n n n n n i i i i distinct distinct from

The irthday problem n distinct n i i distinct from i i distinct i distinct from i i distinct 365 i + 365 365 364 365 n + n distinct 365 365 365 365 364 365 n + n not distinct 365 365 365

The irthday problem For 23 people the probability of sharing a birthday is above 0.5! http://en.wikipedia.org/wiki/irthday_problem

Independence Two events and are independent if and only if In other words and This is an important simplifying assumption for modeling e.g. Toothache and Weather can be assumed to be independent re two mutually exclusive events independent? No but for mutually exclusive events we have +

Independence Two events and are independent if and only if In other words and This is an important simplifying assumption for modeling e.g. Toothache and Weather can be assumed to be independent Conditional independence: and are conditionally independent given C iff C C C Equivalently: C C or C C

Conditional independence: Example Toothache: boolean variable indicating whether the patient has a toothache Cavity: boolean variable indicating whether the patient has a cavity Catch: whether the dentist s probe catches in the cavity If the patient has a cavity the probability that the probe catches in it doesn't depend on whether he/she has a toothache Catch Toothache Cavity Catch Cavity Therefore Catch is conditionally independent of Toothache given Cavity Likewise Toothache is conditionally independent of Catch given Cavity Toothache Catch Cavity Toothache Cavity Equivalent statement: Toothache Catch Cavity Toothache Cavity Catch Cavity

Conditional independence: Example How many numbers do we need to represent the joint probability table Toothache Cavity Catch? 2 3 7 independent entries Write out the joint distribution using chain rule: Toothache Catch Cavity Cavity Catch Cavity Toothache Catch Cavity Cavity Catch Cavity Toothache Cavity How many numbers do we need to represent these distributions? + 2 + 2 5 independent numbers In most cases the use of conditional independence reduces the size of the representation of the joint distribution from exponential in n to linear in n