Bayesian Networks. Probability Theory and Probabilistic Reasoning. Emma Rollon and Javier Larrosa Q

Size: px
Start display at page:

Download "Bayesian Networks. Probability Theory and Probabilistic Reasoning. Emma Rollon and Javier Larrosa Q"

Transcription

1 Bayesian Networks Probability Theory and Probabilistic Reasoning Emma Rollon and Javier Larrosa Q Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

2 Degrees of belief Given a formula F, propositional logics classifies sentences into one of three categories: Sentences that are implied by F (i.e., F = α) Sentences whose negation are implied by F (i.e., F = α) And all other sentences (i.e., F = α and F = α) It imposes a binary classification: any assignment is either possible or impossible whether it satisfies or contradicts F. Under some circumstances, a finer classification is required: we can assign a degree of belief or probability in [0, 1] to each assignment (a.k.a, world). Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

3 Degrees of belief Eartquake Burglary Alarm Prossible? Pr( ) true true true true.0190 true true false false.0010 true false true true.0560 true false false false.0240 false true true true.1620 false true false false.0180 false false true true.0072 false false false true.7128 Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

4 Probabilistic model A set Z of mutually exclussive outcomes (Z is usually defined as the crossed product of a set of random variables) A probability measure P : Z R that satisfies the following axioms: 0 P(z) 1 for every z Z z Z P(z) = 1 (This is a normalization convention that allows us to directly compare the degrees of belief held by different assignments). The Table mapping each assignment with its probability is known as joint probability distribution. Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

5 Probability of a sentence A sentence α (a.k.a., event) will refer to the set of assignments z that satisfies α. The probability of a sentence α is: Pr(α) = z =α Pr(z) Recall that this expression is known as prior probability. Given two sentences α and β: α β: assignments satisfying α union assignments satisfying β. α β: assignments satisfying α intersection assignments satisfying β. α: assignments not satisfying α. Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

6 Probability of a sentence: example Eartquake Burglary Alarm Pr( ) true true true.0190 true true false.0010 true false true.0560 true false false.0240 false true true.1620 false true false.0180 false false true.0072 false false false.7128 What would be the probability of Earthquake (i.e., Pr(Earthquake)), Burglary (Pr(Burglary)), and not Earthquake (Pr( Earthquake))? Observation The joint probability distribution is usually too large to allow a direct representation as a table. Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

7 Properties of Beliefs Prop1: 0 Pr(α) 1 Prop2: Pr(α) = 1, iff α is a tautology Prop3: Pr(α) = 0, iff α is a contradiction Prop4: Since each assignment must either satisfy α or α and Pr(z) = 1, then Pr(α) + Pr( α) = 1 z Prop5: The assignments satisfying α β can be partition into those satisfying α, those satisfying β, and those satisfying α β. Then, Pr(α β) = Pr(α) + Pr(β) Pr(α β) Recall that this property is known as the inclusion-exclusion principle. Observation Is there any case where Pr(α β) = Pr(α) + Pr(β)? Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

8 Properties of Beliefs: example Eartquake Burglary Alarm Pr( ) true true true.0190 true true false.0010 true false true.0560 true false false.0240 false true true.1620 false true false.0180 false false true.0072 false false false.7128 What would be the probability of Earthquake Burglary? Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

9 Updating beliefs Suppose that we know that the Alarm has triggered (i.e., Alarm = true). This new piece of information is called evidence. We have to update the beliefs in Pr( ) into a new state of beliefs Pr( Alarm = true) (i.e., we condition the old table Pr on evidence): Eartquake Burglary Alarm Pr( ) Pr( Alarm) true true true /.2442 true true false true false true /.2442 true false false false true true /.2442 false true false false false true /.2442 false false false Bayes conditioning: Pr(α β) = Pr(α β) Pr(β), when Pr(β) 0 Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

10 Updating beliefs: example Eartquake Burglary Alarm Pr( ) Pr( Alarm) true true true true true false true false true true false false false true true false true false false false true false false false Some changes in beliefs induced by this new evidence: Pr(Burglary) =.2, Pr(Burglary Alarm) =.741 Pr(Earthquake) =.1, Pr(Earthquake Alarm) =.307 Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

11 Beliefs dynamics Let s see how some beliefs would change upon accepting the evidence: Earthquake: Pr(Burglary) =.2, Pr(Burglary Earthquake) =.2 Pr(Alarm) =.2442, Pr(Alarm Earthquake) =.75 Burglary: Pr(Alarm) =.2442, Pr(Alarm Burglary) =.905 Pr(Earthquake) =.1, Pr(Earthquake Burglary) =.1 Alarm and Earthquake: Pr(Burglary) =.2 Pr(Burglary Alarm) =.741 Pr(Burglary Alarm Earthquake) =.253 Alarm and Earthquake: Pr(Burglary Alarm Earthquake) =.957 Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

12 Event independence An event α is independent of event β (noted α β) iff Pr(α β) = Pr(α) Equivalently, Pr(α β) = Pr(α)Pr(β) Example: Pr(Burglary) =.2, Pr(Burglary Earthquake) =.2 (i.e., Burglary Earthquake) Pr(Earthquake) =.1, Pr(Earthquake Burglary) =.1 An event α is conditional independent of event β given event δ (noted α β δ) iff Pr(α β δ) = Pr(α δ) Example: Pr(Burglary Earthquake Alarm) =.253 (i.e., Burglary Earthquake Alarm) Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

13 Event independence: comments Some warnings on conditional independence: α β δ does not imply α β α β does not imply α β δ However, α β iff β α Difference between independence and logical disjointness: Logical disjointness (mutual exclusiveness) iff the models of α and the models of β don t intersect. Independence depends on the actual beliefs (it is a property of the beliefs). Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

14 Variable independence Variables A and B are independent (noted A B) if P(a b) = P(a) for all possible instantiation a, b of A, B. Variables A and B are independent given C (noted A B C) if P(a b, c) = P(a c) for all possible instantiation a, b, c of A, B, C. Observation Variable independence is stronger than event independence. Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

15 Summary of properties of beliefs Prior probability: Pr(α) = z =α Pr(z) Inclusion-exclusion principle: Bayes conditioning: Pr(α β) = Pr(α) + Pr(β) Pr(α β) Pr(α β) = Pr(α β) Pr(β) Product rule: Pr(α β)pr(β) = Pr(α β) Chain rule: Pr(α β 1... β n ) = Pr(α β 1... β n )Pr(β 1 β 2... β n )... Pr(β n ) Bayes rule: Pr(α β)pr(β) = Pr(β α)pr(α) Low of total probability: Given that β 1,..., β n are mutually exclusive: n n Pr(α) = Pr(α β i ) = Pr(α β i )Pr(β i ) i=1 Emma Rollon and Javier Larrosa Bayesian Networks Q / 26 i=1

16 Example: Diagnosis We want to diagnose whether a patient has a certain disease (e.g., cancer) or not. To do that, a test on that disease is available. Let the proposition variable C stand for disease (i.e., C = true (noted c) means that the patience has the disease, and C = false (noted c) means the patience is healthy), and the proposition variable T stand for test (t when the test comes out positive, t otherwise). If we know that: P(c) =.01, P(t c) =.9, P(t c) =.2, P( c) = P( t c) = P( t c) = Observe that... P(C T ) = P(T C)P(C) P(C T ) = P(C) P(c t) =.04 P(c t) =.001 P(t) =.207 P(T C) P(T ) Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

17 Example: Diagnosis with two independent tests Now, we have available two tests (T 1 and T 2 ). We know that: P(c) =.01, P(t i c) =.9, P(t i c) =.2, P( c) = P( t i c) = P( t i c) = and that T 1 and T 2 are conditionally independent given C (i.e., T 1 T 2 C). Observe that... P(C, T 1, T 2 ) = P(T 1 C, T 2 )P(T 2 C)P(C) = P(T 1 C)P(T 2 C)P(C) P(c t 1 ) =.04 P(c t 1, t 2 ) =.16 P(t 1 ) =.207 P(t 1 t 2 ) =.23 Observation Note that T 1 and T 2 are NOT independent. Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

18 Example: Conditional independence We have the weather forecast of three consecutive days (we know whether those days were sunny or not). Let the propositional variables S 1, S 2, and S 3 refer to each of those days, and let S i = s when that day was sunny and S i = s otherwise. We know that: P(s 1 ) =.9, P(s i s i 1 ) =.95, P(s i s i 1 ) =.6 P( s 1 ) =.1, P( s i s i 1 ) =.005, P( s i s i 1 ) =.4 and that S 1 S 3 S 2. Observe that... P(S 1, S 2, S 3 ) = P(S 3 S 2 )P(S 2 S 1 )P(S 1 ) P(s 1 ) =.90, P(s 2 ) =.91, P(s 3 ) =.92 P(s 1 s 3 ) =.91, P(s 3 s 1 ) =.93 Observation Note that S1 and S 3 are NOT independent. Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

19 Example: Explaining away I am happy (H) when it is sunny (S) or when I get a salary raise (R). We know that: P(s) =.7 P(r) =.01 P(h s, r) = 1 P(h s, r) =.9 P(h s, r) =.7 P(h s, r) =.1 and we assume S R. P(H, S, R) = P(H S, R)P(S, R) = P(H S, R)P(S)P(R) P(r h) =.018 P(r h, s) =.014 Observation Note that R and S are NOT independent given H Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

20 Most usual queries Let X = E Q Z, where E are the evidence variables, e are the observed values, Q are the variables of interest, Z are the rest of the variables. Probability of Evidence (Q = ): Prior Marginals (E = ): Posterior Marginals (E ): Pr(e) = Z Pr(e, Z) Pr(Q) = Z Pr(Q, Z) Pr(Q e) = Z Pr(Q, Z e) Most Probable Explanation (MPE) (E, Z = ): q = arg max Q Pr(Q e) Maximum a Posteriori Probability (MAP): q = arg max Q Pr(Q e) = arg max Q Z Pr(Q, Z e) Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

21 Most usual queries: Example Recall our Earthquake, Burglary and Alarm example, with joint probability: Eartquake (T) Burglary (B) Alarm (A) Pr( ) true true true.0190 true true false.0010 true false true.0560 true false false.0240 false true true.1620 false true false.0180 false false true.0072 false false false.7128 Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

22 Most usual queries: Example (Probability of Evidence) Eartquake (T) Burglary (B) Alarm (A) Pr( ) true true true.0190 true true false.0010 true false true.0560 true false false.0240 false true true.1620 false true false.0180 false false true.0072 false false false.7128 Pr(A = true) = T,B Pr(T, B, A = true) Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

23 Most usual queries: Example (Prior Marginals) Eartquake (T) Burglary (B) Alarm (A) Pr( ) true true true.0190 true true false.0010 true false true.0560 true false false.0240 false true true.1620 false true false.0180 false false true.0072 false false false.7128 Pr(T ) = A,B Pr(T, B, A) Eartquake (T) true false Pr( ) Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

24 Most usual queries: Example (Posterior Marginals) Eartquake (T) Burglary (B) Alarm (A) Pr( ) true true true.0190 true true false.0010 true false true.0560 true false false.0240 false true true.1620 false true false.0180 false false true.0072 false false false.7128 Pr(T A = true) = B Pr(T, B A = true) = B Pr(T,B,A=true) Pr(A=true) Eartquake (T) true false Pr( A = true) Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

25 Most usual queries: Example (Most Probable Explanation) Eartquake (T) Burglary (B) Alarm (A) Pr( ) true true true.0190 true true false.0010 true false true.0560 true false false.0240 false true true.1620 false true false.0180 false false true.0072 false false false.7128 arg max T,B Pr(T, B A = true) = arg max T,B Pr(T,B,A=true) Pr(A=true) = arg max T,B Pr(T, B, A = true) Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

26 Most usual queries: Example (Maximum a Posteriori Probability) Eartquake (T) Burglary (B) Alarm (A) Pr( ) true true true.0190 true true false.0010 true false true.0560 true false false.0240 false true true.1620 false true false.0180 false false true.0072 false false false.7128 arg max T Pr(T A = true) = arg max T B Pr(T, B A = true) = Pr(T,B,A=true) arg max T B Pr(A=true) = arg max T B Pr(T, B, A = true) Emma Rollon and Javier Larrosa Bayesian Networks Q / 26

Refresher on Probability Theory

Refresher on Probability Theory Much of this material is adapted from Chapters 2 and 3 of Darwiche s book January 16, 2014 1 Preliminaries 2 Degrees of Belief 3 Independence 4 Other Important Properties 5 Wrap-up Primitives The following

More information

Probability Calculus. Chapter From Propositional to Graded Beliefs

Probability Calculus. Chapter From Propositional to Graded Beliefs Chapter 2 Probability Calculus Our purpose in this chapter is to introduce probability calculus and then show how it can be used to represent uncertain beliefs, and then change them in the face of new

More information

Reasoning with Bayesian Networks

Reasoning with Bayesian Networks Reasoning with Lecture 1: Probability Calculus, NICTA and ANU Reasoning with Overview of the Course Probability calculus, Bayesian networks Inference by variable elimination, factor elimination, conditioning

More information

Logic and Bayesian Networks

Logic and Bayesian Networks Logic and Part 1: and Jinbo Huang Jinbo Huang and 1/ 31 What This Course Is About Probabilistic reasoning with Bayesian networks Reasoning by logical encoding and compilation Jinbo Huang and 2/ 31 Probabilities

More information

Bayesian Networks. Exact Inference by Variable Elimination. Emma Rollon and Javier Larrosa Q

Bayesian Networks. Exact Inference by Variable Elimination. Emma Rollon and Javier Larrosa Q Bayesian Networks Exact Inference by Variable Elimination Emma Rollon and Javier Larrosa Q1-2015-2016 Emma Rollon and Javier Larrosa Bayesian Networks Q1-2015-2016 1 / 25 Recall the most usual queries

More information

Probability Calculus. p.1

Probability Calculus. p.1 Probability Calculus p.1 Joint probability distribution A state of belief : assign a degree of belief or probability to each world 0 Pr ω 1 ω Ω sentence > event world > elementary event Pr ω ω α 1 Pr ω

More information

Introduction to Probabilistic Reasoning. Image credit: NASA. Assignment

Introduction to Probabilistic Reasoning. Image credit: NASA. Assignment Introduction to Probabilistic Reasoning Brian C. Williams 16.410/16.413 November 17 th, 2010 11/17/10 copyright Brian Williams, 2005-10 1 Brian C. Williams, copyright 2000-09 Image credit: NASA. Assignment

More information

Bayesian Reasoning. Adapted from slides by Tim Finin and Marie desjardins.

Bayesian Reasoning. Adapted from slides by Tim Finin and Marie desjardins. Bayesian Reasoning Adapted from slides by Tim Finin and Marie desjardins. 1 Outline Probability theory Bayesian inference From the joint distribution Using independence/factoring From sources of evidence

More information

STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas. Propositional Logic and Probability Theory: Review

STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas. Propositional Logic and Probability Theory: Review STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas Propositional Logic and Probability Theory: Review Logic Logics are formal languages for representing information such that

More information

Quantifying uncertainty & Bayesian networks

Quantifying uncertainty & Bayesian networks Quantifying uncertainty & Bayesian networks CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2016 Soleymani Artificial Intelligence: A Modern Approach, 3 rd Edition,

More information

Probability. CS 3793/5233 Artificial Intelligence Probability 1

Probability. CS 3793/5233 Artificial Intelligence Probability 1 CS 3793/5233 Artificial Intelligence 1 Motivation Motivation Random Variables Semantics Dice Example Joint Dist. Ex. Axioms Agents don t have complete knowledge about the world. Agents need to make decisions

More information

Uncertainty. Variables. assigns to each sentence numerical degree of belief between 0 and 1. uncertainty

Uncertainty. Variables. assigns to each sentence numerical degree of belief between 0 and 1. uncertainty Bayes Classification n Uncertainty & robability n Baye's rule n Choosing Hypotheses- Maximum a posteriori n Maximum Likelihood - Baye's concept learning n Maximum Likelihood of real valued function n Bayes

More information

An AI-ish view of Probability, Conditional Probability & Bayes Theorem

An AI-ish view of Probability, Conditional Probability & Bayes Theorem An AI-ish view of Probability, Conditional Probability & Bayes Theorem Review: Uncertainty and Truth Values: a mismatch Let action A t = leave for airport t minutes before flight. Will A 15 get me there

More information

10/18/2017. An AI-ish view of Probability, Conditional Probability & Bayes Theorem. Making decisions under uncertainty.

10/18/2017. An AI-ish view of Probability, Conditional Probability & Bayes Theorem. Making decisions under uncertainty. An AI-ish view of Probability, Conditional Probability & Bayes Theorem Review: Uncertainty and Truth Values: a mismatch Let action A t = leave for airport t minutes before flight. Will A 15 get me there

More information

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty Lecture 10: Introduction to reasoning under uncertainty Introduction to reasoning under uncertainty Review of probability Axioms and inference Conditional probability Probability distributions COMP-424,

More information

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)

More information

Quantifying Uncertainty & Probabilistic Reasoning. Abdulla AlKhenji Khaled AlEmadi Mohammed AlAnsari

Quantifying Uncertainty & Probabilistic Reasoning. Abdulla AlKhenji Khaled AlEmadi Mohammed AlAnsari Quantifying Uncertainty & Probabilistic Reasoning Abdulla AlKhenji Khaled AlEmadi Mohammed AlAnsari Outline Previous Implementations What is Uncertainty? Acting Under Uncertainty Rational Decisions Basic

More information

Uncertainty. Logic and Uncertainty. Russell & Norvig. Readings: Chapter 13. One problem with logical-agent approaches: C:145 Artificial

Uncertainty. Logic and Uncertainty. Russell & Norvig. Readings: Chapter 13. One problem with logical-agent approaches: C:145 Artificial C:145 Artificial Intelligence@ Uncertainty Readings: Chapter 13 Russell & Norvig. Artificial Intelligence p.1/43 Logic and Uncertainty One problem with logical-agent approaches: Agents almost never have

More information

Uncertainty. Introduction to Artificial Intelligence CS 151 Lecture 2 April 1, CS151, Spring 2004

Uncertainty. Introduction to Artificial Intelligence CS 151 Lecture 2 April 1, CS151, Spring 2004 Uncertainty Introduction to Artificial Intelligence CS 151 Lecture 2 April 1, 2004 Administration PA 1 will be handed out today. There will be a MATLAB tutorial tomorrow, Friday, April 2 in AP&M 4882 at

More information

Directed Graphical Models

Directed Graphical Models CS 2750: Machine Learning Directed Graphical Models Prof. Adriana Kovashka University of Pittsburgh March 28, 2017 Graphical Models If no assumption of independence is made, must estimate an exponential

More information

Probabilistic Classification

Probabilistic Classification Bayesian Networks Probabilistic Classification Goal: Gather Labeled Training Data Build/Learn a Probability Model Use the model to infer class labels for unlabeled data points Example: Spam Filtering...

More information

Uncertainty (Chapter 13, Russell & Norvig) Introduction to Artificial Intelligence CS 150 Lecture 14

Uncertainty (Chapter 13, Russell & Norvig) Introduction to Artificial Intelligence CS 150 Lecture 14 Uncertainty (Chapter 13, Russell & Norvig) Introduction to Artificial Intelligence CS 150 Lecture 14 Administration Last Programming assignment will be handed out later this week. I am doing probability

More information

Y. Xiang, Inference with Uncertain Knowledge 1

Y. Xiang, Inference with Uncertain Knowledge 1 Inference with Uncertain Knowledge Objectives Why must agent use uncertain knowledge? Fundamentals of Bayesian probability Inference with full joint distributions Inference with Bayes rule Bayesian networks

More information

Uncertain Reasoning. Environment Description. Configurations. Models. Bayesian Networks

Uncertain Reasoning. Environment Description. Configurations. Models. Bayesian Networks Bayesian Networks A. Objectives 1. Basics on Bayesian probability theory 2. Belief updating using JPD 3. Basics on graphs 4. Bayesian networks 5. Acquisition of Bayesian networks 6. Local computation and

More information

Bayesian Learning. Examples. Conditional Probability. Two Roles for Bayesian Methods. Prior Probability and Random Variables. The Chain Rule P (B)

Bayesian Learning. Examples. Conditional Probability. Two Roles for Bayesian Methods. Prior Probability and Random Variables. The Chain Rule P (B) Examples My mood can take 2 possible values: happy, sad. The weather can take 3 possible vales: sunny, rainy, cloudy My friends know me pretty well and say that: P(Mood=happy Weather=rainy) = 0.25 P(Mood=happy

More information

Ch.6 Uncertain Knowledge. Logic and Uncertainty. Representation. One problem with logical approaches: Department of Computer Science

Ch.6 Uncertain Knowledge. Logic and Uncertainty. Representation. One problem with logical approaches: Department of Computer Science Ch.6 Uncertain Knowledge Representation Hantao Zhang http://www.cs.uiowa.edu/ hzhang/c145 The University of Iowa Department of Computer Science Artificial Intelligence p.1/39 Logic and Uncertainty One

More information

Bayesian belief networks

Bayesian belief networks CS 2001 Lecture 1 Bayesian belief networks Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square 4-8845 Milos research interests Artificial Intelligence Planning, reasoning and optimization in the presence

More information

Basic Probabilistic Reasoning SEG

Basic Probabilistic Reasoning SEG Basic Probabilistic Reasoning SEG 7450 1 Introduction Reasoning under uncertainty using probability theory Dealing with uncertainty is one of the main advantages of an expert system over a simple decision

More information

Axioms of Probability? Notation. Bayesian Networks. Bayesian Networks. Today we ll introduce Bayesian Networks.

Axioms of Probability? Notation. Bayesian Networks. Bayesian Networks. Today we ll introduce Bayesian Networks. Bayesian Networks Today we ll introduce Bayesian Networks. This material is covered in chapters 13 and 14. Chapter 13 gives basic background on probability and Chapter 14 talks about Bayesian Networks.

More information

Preliminaries Bayesian Networks Graphoid Axioms d-separation Wrap-up. Bayesian Networks. Brandon Malone

Preliminaries Bayesian Networks Graphoid Axioms d-separation Wrap-up. Bayesian Networks. Brandon Malone Preliminaries Graphoid Axioms d-separation Wrap-up Much of this material is adapted from Chapter 4 of Darwiche s book January 23, 2014 Preliminaries Graphoid Axioms d-separation Wrap-up 1 Preliminaries

More information

Pengju

Pengju Introduction to AI Chapter13 Uncertainty Pengju Ren@IAIR Outline Uncertainty Probability Syntax and Semantics Inference Independence and Bayes Rule Example: Car diagnosis Wumpus World Environment Squares

More information

Uncertainty. Chapter 13

Uncertainty. Chapter 13 Uncertainty Chapter 13 Outline Uncertainty Probability Syntax and Semantics Inference Independence and Bayes Rule Uncertainty Let s say you want to get to the airport in time for a flight. Let action A

More information

Artificial Intelligence Programming Probability

Artificial Intelligence Programming Probability Artificial Intelligence Programming Probability Chris Brooks Department of Computer Science University of San Francisco Department of Computer Science University of San Francisco p.1/?? 13-0: Uncertainty

More information

Introduction to Artificial Intelligence. Unit # 11

Introduction to Artificial Intelligence. Unit # 11 Introduction to Artificial Intelligence Unit # 11 1 Course Outline Overview of Artificial Intelligence State Space Representation Search Techniques Machine Learning Logic Probabilistic Reasoning/Bayesian

More information

Modeling and reasoning with uncertainty

Modeling and reasoning with uncertainty CS 2710 Foundations of AI Lecture 18 Modeling and reasoning with uncertainty Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square KB systems. Medical example. We want to build a KB system for the diagnosis

More information

Lecture 8. Probabilistic Reasoning CS 486/686 May 25, 2006

Lecture 8. Probabilistic Reasoning CS 486/686 May 25, 2006 Lecture 8 Probabilistic Reasoning CS 486/686 May 25, 2006 Outline Review probabilistic inference, independence and conditional independence Bayesian networks What are they What do they mean How do we create

More information

CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-II

CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-II CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-II 1 Bayes Rule Example Disease {malaria, cold, flu}; Symptom = fever Must compute Pr(D fever) to prescribe treatment Why not assess

More information

UNCERTAINTY. In which we see what an agent should do when not all is crystal-clear.

UNCERTAINTY. In which we see what an agent should do when not all is crystal-clear. UNCERTAINTY In which we see what an agent should do when not all is crystal-clear. Outline Uncertainty Probabilistic Theory Axioms of Probability Probabilistic Reasoning Independency Bayes Rule Summary

More information

Introduction to Bayesian Learning

Introduction to Bayesian Learning Course Information Introduction Introduction to Bayesian Learning Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Apprendimento Automatico: Fondamenti - A.A. 2016/2017 Outline

More information

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2016/2017 Lesson 13 24 march 2017 Reasoning with Bayesian Networks Naïve Bayesian Systems...2 Example

More information

COMP5211 Lecture Note on Reasoning under Uncertainty

COMP5211 Lecture Note on Reasoning under Uncertainty COMP5211 Lecture Note on Reasoning under Uncertainty Fangzhen Lin Department of Computer Science and Engineering Hong Kong University of Science and Technology Fangzhen Lin (HKUST) Uncertainty 1 / 33 Uncertainty

More information

Computer Science CPSC 322. Lecture 18 Marginalization, Conditioning

Computer Science CPSC 322. Lecture 18 Marginalization, Conditioning Computer Science CPSC 322 Lecture 18 Marginalization, Conditioning Lecture Overview Recap Lecture 17 Joint Probability Distribution, Marginalization Conditioning Inference by Enumeration Bayes Rule, Chain

More information

Defining Things in Terms of Joint Probability Distribution. Today s Lecture. Lecture 17: Uncertainty 2. Victor R. Lesser

Defining Things in Terms of Joint Probability Distribution. Today s Lecture. Lecture 17: Uncertainty 2. Victor R. Lesser Lecture 17: Uncertainty 2 Victor R. Lesser CMPSCI 683 Fall 2010 Today s Lecture How belief networks can be a Knowledge Base for probabilistic knowledge. How to construct a belief network. How to answer

More information

Resolution or modus ponens are exact there is no possibility of mistake if the rules are followed exactly.

Resolution or modus ponens are exact there is no possibility of mistake if the rules are followed exactly. THE WEAKEST LINK Resolution or modus ponens are exact there is no possibility of mistake if the rules are followed exactly. These methods of inference (also known as deductive methods) require that information

More information

Motivation. Bayesian Networks in Epistemology and Philosophy of Science Lecture. Overview. Organizational Issues

Motivation. Bayesian Networks in Epistemology and Philosophy of Science Lecture. Overview. Organizational Issues Bayesian Networks in Epistemology and Philosophy of Science Lecture 1: Bayesian Networks Center for Logic and Philosophy of Science Tilburg University, The Netherlands Formal Epistemology Course Northern

More information

Artificial Intelligence Uncertainty

Artificial Intelligence Uncertainty Artificial Intelligence Uncertainty Ch. 13 Uncertainty Let action A t = leave for airport t minutes before flight Will A t get me there on time? A 25, A 60, A 3600 Uncertainty: partial observability (road

More information

CS 188: Artificial Intelligence. Bayes Nets

CS 188: Artificial Intelligence. Bayes Nets CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew

More information

Probability. Lecture Notes. Adolfo J. Rumbos

Probability. Lecture Notes. Adolfo J. Rumbos Probability Lecture Notes Adolfo J. Rumbos October 20, 204 2 Contents Introduction 5. An example from statistical inference................ 5 2 Probability Spaces 9 2. Sample Spaces and σ fields.....................

More information

This lecture. Reading. Conditional Independence Bayesian (Belief) Networks: Syntax and semantics. Chapter CS151, Spring 2004

This lecture. Reading. Conditional Independence Bayesian (Belief) Networks: Syntax and semantics. Chapter CS151, Spring 2004 This lecture Conditional Independence Bayesian (Belief) Networks: Syntax and semantics Reading Chapter 14.1-14.2 Propositions and Random Variables Letting A refer to a proposition which may either be true

More information

Probabilistic Representation and Reasoning

Probabilistic Representation and Reasoning Probabilistic Representation and Reasoning Alessandro Panella Department of Computer Science University of Illinois at Chicago May 4, 2010 Alessandro Panella (CS Dept. - UIC) Probabilistic Representation

More information

Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule

Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule Alan Mackworth UBC CS 322 Uncertainty 2 March 13, 2013 Textbook 6.1.3 Lecture Overview Recap: Probability & Possible World Semantics

More information

Web-Mining Agents Data Mining

Web-Mining Agents Data Mining Web-Mining Agents Data Mining Prof. Dr. Ralf Möller Dr. Özgür L. Özçep Universität zu Lübeck Institut für Informationssysteme Tanya Braun (Übungen) 2 Uncertainty AIMA Chapter 13 3 Outline Agents Uncertainty

More information

Probabilistic representation and reasoning

Probabilistic representation and reasoning Probabilistic representation and reasoning Applied artificial intelligence (EDA132) Lecture 09 2017-02-15 Elin A. Topp Material based on course book, chapter 13, 14.1-3 1 Show time! Two boxes of chocolates,

More information

Probabilistic Robotics

Probabilistic Robotics Probabilistic Robotics Overview of probability, Representing uncertainty Propagation of uncertainty, Bayes Rule Application to Localization and Mapping Slides from Autonomous Robots (Siegwart and Nourbaksh),

More information

CS 484 Data Mining. Classification 7. Some slides are from Professor Padhraic Smyth at UC Irvine

CS 484 Data Mining. Classification 7. Some slides are from Professor Padhraic Smyth at UC Irvine CS 484 Data Mining Classification 7 Some slides are from Professor Padhraic Smyth at UC Irvine Bayesian Belief networks Conditional independence assumption of Naïve Bayes classifier is too strong. Allows

More information

Probabilistic representation and reasoning

Probabilistic representation and reasoning Probabilistic representation and reasoning Applied artificial intelligence (EDAF70) Lecture 04 2019-02-01 Elin A. Topp Material based on course book, chapter 13, 14.1-3 1 Show time! Two boxes of chocolates,

More information

Uncertainty and Belief Networks. Introduction to Artificial Intelligence CS 151 Lecture 1 continued Ok, Lecture 2!

Uncertainty and Belief Networks. Introduction to Artificial Intelligence CS 151 Lecture 1 continued Ok, Lecture 2! Uncertainty and Belief Networks Introduction to Artificial Intelligence CS 151 Lecture 1 continued Ok, Lecture 2! This lecture Conditional Independence Bayesian (Belief) Networks: Syntax and semantics

More information

Artificial Intelligence CS 6364

Artificial Intelligence CS 6364 Artificial Intelligence CS 6364 rofessor Dan Moldovan Section 12 robabilistic Reasoning Acting under uncertainty Logical agents assume propositions are - True - False - Unknown acting under uncertainty

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning CS4375 --- Fall 2018 Bayesian a Learning Reading: Sections 13.1-13.6, 20.1-20.2, R&N Sections 6.1-6.3, 6.7, 6.9, Mitchell 1 Uncertainty Most real-world problems deal with

More information

10/15/2015 A FAST REVIEW OF DISCRETE PROBABILITY (PART 2) Probability, Conditional Probability & Bayes Rule. Discrete random variables

10/15/2015 A FAST REVIEW OF DISCRETE PROBABILITY (PART 2) Probability, Conditional Probability & Bayes Rule. Discrete random variables Probability, Conditional Probability & Bayes Rule A FAST REVIEW OF DISCRETE PROBABILITY (PART 2) 2 Discrete random variables A random variable can take on one of a set of different values, each with an

More information

MATH 556: PROBABILITY PRIMER

MATH 556: PROBABILITY PRIMER MATH 6: PROBABILITY PRIMER 1 DEFINITIONS, TERMINOLOGY, NOTATION 1.1 EVENTS AND THE SAMPLE SPACE Definition 1.1 An experiment is a one-off or repeatable process or procedure for which (a there is a well-defined

More information

Probability and Uncertainty. Bayesian Networks

Probability and Uncertainty. Bayesian Networks Probability and Uncertainty Bayesian Networks First Lecture Today (Tue 28 Jun) Review Chapters 8.1-8.5, 9.1-9.2 (optional 9.5) Second Lecture Today (Tue 28 Jun) Read Chapters 13, & 14.1-14.5 Next Lecture

More information

Introduction to Machine Learning

Introduction to Machine Learning Uncertainty Introduction to Machine Learning CS4375 --- Fall 2018 a Bayesian Learning Reading: Sections 13.1-13.6, 20.1-20.2, R&N Sections 6.1-6.3, 6.7, 6.9, Mitchell Most real-world problems deal with

More information

CMPT Machine Learning. Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th

CMPT Machine Learning. Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th CMPT 882 - Machine Learning Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th Stephen Fagan sfagan@sfu.ca Overview: Introduction - Who was Bayes? - Bayesian Statistics Versus Classical Statistics

More information

Uncertainty and Bayesian Networks

Uncertainty and Bayesian Networks Uncertainty and Bayesian Networks Tutorial 3 Tutorial 3 1 Outline Uncertainty Probability Syntax and Semantics for Uncertainty Inference Independence and Bayes Rule Syntax and Semantics for Bayesian Networks

More information

Recitation 2: Probability

Recitation 2: Probability Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions

More information

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber Data Modeling & Analysis Techniques Probability & Statistics Manfred Huber 2017 1 Probability and Statistics Probability and statistics are often used interchangeably but are different, related fields

More information

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018 Bayesian networks CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018 Soleymani Slides have been adopted from Klein and Abdeel, CS188, UC Berkeley. Outline Probability

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Bayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014

Bayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014 Bayes Formula MATH 07: Finite Mathematics University of Louisville March 26, 204 Test Accuracy Conditional reversal 2 / 5 A motivating question A rare disease occurs in out of every 0,000 people. A test

More information

Uncertainty. Russell & Norvig Chapter 13.

Uncertainty. Russell & Norvig Chapter 13. Uncertainty Russell & Norvig Chapter 13 http://toonut.com/wp-content/uploads/2011/12/69wp.jpg Uncertainty Let A t be the action of leaving for the airport t minutes before your flight Will A t get you

More information

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 Plan for today and next week Today and next time: Bayesian networks (Bishop Sec. 8.1) Conditional

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

CS 188: Artificial Intelligence. Our Status in CS188

CS 188: Artificial Intelligence. Our Status in CS188 CS 188: Artificial Intelligence Probability Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein. 1 Our Status in CS188 We re done with Part I Search and Planning! Part II: Probabilistic Reasoning

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information

Behavioral Data Mining. Lecture 2

Behavioral Data Mining. Lecture 2 Behavioral Data Mining Lecture 2 Autonomy Corp Bayes Theorem Bayes Theorem P(A B) = probability of A given that B is true. P(A B) = P(B A)P(A) P(B) In practice we are most interested in dealing with events

More information

STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas. Markov networks: Representation

STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas. Markov networks: Representation STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas Markov networks: Representation Markov networks: Undirected Graphical models Model the following distribution using a Bayesian

More information

Sample Space: Specify all possible outcomes from an experiment. Event: Specify a particular outcome or combination of outcomes.

Sample Space: Specify all possible outcomes from an experiment. Event: Specify a particular outcome or combination of outcomes. Chapter 2 Introduction to Probability 2.1 Probability Model Probability concerns about the chance of observing certain outcome resulting from an experiment. However, since chance is an abstraction of something

More information

UNCERTAIN KNOWLEDGE AND REASONING

UNCERTAIN KNOWLEDGE AND REASONING QUESTION BANK DEPARTMENT: CSE SEMESTER: VI SUBJECT CODE / Name: CS2351 ARTIFICIAL INTELLIGENCE UNIT IV PART -A (2 Marks) UNCERTAIN KNOWLEDGE AND REASONING 1. List down two applications of temporal probabilistic

More information

Why Probability? It's the right way to look at the world.

Why Probability? It's the right way to look at the world. Probability Why Probability? It's the right way to look at the world. Discrete Random Variables We denote discrete random variables with capital letters. A boolean random variable may be either true or

More information

FIRST ORDER LOGIC AND PROBABILISTIC INFERENCING. ECE457 Applied Artificial Intelligence Page 1

FIRST ORDER LOGIC AND PROBABILISTIC INFERENCING. ECE457 Applied Artificial Intelligence Page 1 FIRST ORDER LOGIC AND PROBABILISTIC INFERENCING ECE457 Applied Artificial Intelligence Page 1 Resolution Recall from Propositional Logic (αβ), ( βγ) (α γ) Resolution rule is both sound and complete Idea

More information

Be able to define the following terms and answer basic questions about them:

Be able to define the following terms and answer basic questions about them: CS440/ECE448 Section Q Fall 2017 Final Review Be able to define the following terms and answer basic questions about them: Probability o Random variables, axioms of probability o Joint, marginal, conditional

More information

Uncertainty in the World. Representing Uncertainty. Uncertainty in the World and our Models. Uncertainty

Uncertainty in the World. Representing Uncertainty. Uncertainty in the World and our Models. Uncertainty Uncertainty in the World Representing Uncertainty Chapter 13 An agent can often be uncertain about the state of the world/domain since there is often ambiguity and uncertainty Plausible/probabilistic inference

More information

Lecture 9: Naive Bayes, SVM, Kernels. Saravanan Thirumuruganathan

Lecture 9: Naive Bayes, SVM, Kernels. Saravanan Thirumuruganathan Lecture 9: Naive Bayes, SVM, Kernels Instructor: Outline 1 Probability basics 2 Probabilistic Interpretation of Classification 3 Bayesian Classifiers, Naive Bayes 4 Support Vector Machines Probability

More information

Rapid Introduction to Machine Learning/ Deep Learning

Rapid Introduction to Machine Learning/ Deep Learning Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian

More information

Topics. Bayesian Learning. What is Bayesian Learning? Objectives for Bayesian Learning

Topics. Bayesian Learning. What is Bayesian Learning? Objectives for Bayesian Learning Topics Bayesian Learning Sattiraju Prabhakar CS898O: ML Wichita State University Objectives for Bayesian Learning Bayes Theorem and MAP Bayes Optimal Classifier Naïve Bayes Classifier An Example Classifying

More information

Reasoning Under Uncertainty: Conditional Probability

Reasoning Under Uncertainty: Conditional Probability Reasoning Under Uncertainty: Conditional Probability CPSC 322 Uncertainty 2 Textbook 6.1 Reasoning Under Uncertainty: Conditional Probability CPSC 322 Uncertainty 2, Slide 1 Lecture Overview 1 Recap 2

More information

COMP9414: Artificial Intelligence Reasoning Under Uncertainty

COMP9414: Artificial Intelligence Reasoning Under Uncertainty COMP9414, Monday 16 April, 2012 Reasoning Under Uncertainty 2 COMP9414: Artificial Intelligence Reasoning Under Uncertainty Overview Problems with Logical Approach What Do the Numbers Mean? Wayne Wobcke

More information

Uncertainty. Outline

Uncertainty. Outline Uncertainty Chapter 13 Outline Uncertainty Probability Syntax and Semantics Inference Independence and Bayes' Rule 1 Uncertainty Let action A t = leave for airport t minutes before flight Will A t get

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Chapter 8&9: Classification: Part 3 Instructor: Yizhou Sun yzsun@ccs.neu.edu March 12, 2013 Midterm Report Grade Distribution 90-100 10 80-89 16 70-79 8 60-69 4

More information

Bayesian Networks Representation

Bayesian Networks Representation Bayesian Networks Representation Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University March 19 th, 2007 Handwriting recognition Character recognition, e.g., kernel SVMs a c z rr r r

More information

Probabilistic Reasoning Systems

Probabilistic Reasoning Systems Probabilistic Reasoning Systems Dr. Richard J. Povinelli Copyright Richard J. Povinelli rev 1.0, 10/7/2001 Page 1 Objectives You should be able to apply belief networks to model a problem with uncertainty.

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Dr Ahmed Rafat Abas Computer Science Dept, Faculty of Computers and Informatics, Zagazig University arabas@zu.edu.eg http://www.arsaliem.faculty.zu.edu.eg/ Uncertainty Chapter 13

More information

Cartesian-product sample spaces and independence

Cartesian-product sample spaces and independence CS 70 Discrete Mathematics for CS Fall 003 Wagner Lecture 4 The final two lectures on probability will cover some basic methods for answering questions about probability spaces. We will apply them to the

More information

Bayesian belief networks. Inference.

Bayesian belief networks. Inference. Lecture 13 Bayesian belief networks. Inference. Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Midterm exam Monday, March 17, 2003 In class Closed book Material covered by Wednesday, March 12 Last

More information

Uncertainty. Chapter 13

Uncertainty. Chapter 13 Uncertainty Chapter 13 Uncertainty Let action A t = leave for airport t minutes before flight Will A t get me there on time? Problems: 1. partial observability (road state, other drivers' plans, noisy

More information

Philosophy 148 Announcements & Such

Philosophy 148 Announcements & Such Branden Fitelson Philosophy 148 Lecture 1 Philosophy 148 Announcements & Such Administrative Stuff Raul s office is 5323 Tolman (not 301 Moses). We have a permanent location for the Tues. section: 206

More information

Bayesian Inference. Definitions from Probability: Naive Bayes Classifiers: Advantages and Disadvantages of Naive Bayes Classifiers:

Bayesian Inference. Definitions from Probability: Naive Bayes Classifiers: Advantages and Disadvantages of Naive Bayes Classifiers: Bayesian Inference The purpose of this document is to review belief networks and naive Bayes classifiers. Definitions from Probability: Belief networks: Naive Bayes Classifiers: Advantages and Disadvantages

More information

Probability. 25 th September lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.)

Probability. 25 th September lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Probability 25 th September 2017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Properties of Probability Methods of Enumeration Conditional Probability Independent

More information

Uncertainty. 22c:145 Artificial Intelligence. Problem of Logic Agents. Foundations of Probability. Axioms of Probability

Uncertainty. 22c:145 Artificial Intelligence. Problem of Logic Agents. Foundations of Probability. Axioms of Probability Problem of Logic Agents 22c:145 Artificial Intelligence Uncertainty Reading: Ch 13. Russell & Norvig Logic-agents almost never have access to the whole truth about their environments. A rational agent

More information