Modeling and reasoning with uncertainty

Similar documents
Bayesian belief networks

Bayesian belief networks. Inference.

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

Knowledge Representation. Propositional logic

Computer Science CPSC 322. Lecture 18 Marginalization, Conditioning

Bayesian Reasoning. Adapted from slides by Tim Finin and Marie desjardins.

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule

Bayesian decision theory. Nuno Vasconcelos ECE Department, UCSD

Knowledge Representation. Propositional logic.

Uncertainty. Chapter 13

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

An AI-ish view of Probability, Conditional Probability & Bayes Theorem

10/18/2017. An AI-ish view of Probability, Conditional Probability & Bayes Theorem. Making decisions under uncertainty.

Probabilistic Reasoning. (Mostly using Bayesian Networks)

PROBABILISTIC REASONING SYSTEMS

Uncertainty and Bayesian Networks

Pengju XJTU 2016

Why Probability? It's the right way to look at the world.

Basic Probabilistic Reasoning SEG

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Random Variables. A random variable is some aspect of the world about which we (may) have uncertainty

Probabilistic Robotics

Artificial Intelligence CS 6364

Quantifying uncertainty & Bayesian networks

Y. Xiang, Inference with Uncertain Knowledge 1

Uncertainty. Logic and Uncertainty. Russell & Norvig. Readings: Chapter 13. One problem with logical-agent approaches: C:145 Artificial

CS 188: Artificial Intelligence Fall 2011

Artificial Intelligence

CS 188: Artificial Intelligence. Our Status in CS188

Probabilistic Reasoning

Discrete Random Variables

Uncertainty. Introduction to Artificial Intelligence CS 151 Lecture 2 April 1, CS151, Spring 2004

Reasoning about uncertainty

Probabilistic Reasoning

CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-II

Artificial Intelligence Uncertainty

Uncertainty. Variables. assigns to each sentence numerical degree of belief between 0 and 1. uncertainty

Pengju

Where are we in CS 440?

CSE 473: Artificial Intelligence

13.4 INDEPENDENCE. 494 Chapter 13. Quantifying Uncertainty

10/15/2015 A FAST REVIEW OF DISCRETE PROBABILITY (PART 2) Probability, Conditional Probability & Bayes Rule. Discrete random variables

CS 188: Artificial Intelligence Fall 2009

UNCERTAINTY. In which we see what an agent should do when not all is crystal-clear.

Course Introduction. Probabilistic Modelling and Reasoning. Relationships between courses. Dealing with Uncertainty. Chris Williams.

Where are we in CS 440?

n How to represent uncertainty in knowledge? n Which action to choose under uncertainty? q Assume the car does not have a flat tire

CS188 Outline. CS 188: Artificial Intelligence. Today. Inference in Ghostbusters. Probability. We re done with Part I: Search and Planning!

Uncertainty. Outline

Introduction to Artificial Intelligence. Unit # 11

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Uncertainty. Russell & Norvig Chapter 13.

Ch.6 Uncertain Knowledge. Logic and Uncertainty. Representation. One problem with logical approaches: Department of Computer Science

Fusion in simple models

Our Status in CSE 5522

PROBABILITY. Inference: constraint propagation

Probability Hal Daumé III. Computer Science University of Maryland CS 421: Introduction to Artificial Intelligence 27 Mar 2012

Our Status. We re done with Part I Search and Planning!

Uncertainty. Chapter 13

Artificial Intelligence Programming Probability

In today s lecture. Conditional probability and independence. COSC343: Artificial Intelligence. Curse of dimensionality.

COMP5211 Lecture Note on Reasoning under Uncertainty

CS188 Outline. We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning. Part III: Machine Learning

Probabilistic Reasoning in AI COMP-526

Reasoning Under Uncertainty: Conditional Probability

Introduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013

PROBABILITY AND INFERENCE

Uncertain Knowledge and Bayes Rule. George Konidaris

Probabilistic Robotics. Slides from Autonomous Robots (Siegwart and Nourbaksh), Chapter 5 Probabilistic Robotics (S. Thurn et al.

Reasoning under Uncertainty: Intro to Probability

Discrete Probability and State Estimation

Probabilistic Reasoning. KR Chowdhary, Professor, Department of Computer Science & Engineering, MBM Engineering College, JNV University, Jodhpur,

Reasoning with Uncertainty

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

Bayesian Networks and Decision Graphs

Uncertainty. Chapter 13. Chapter 13 1

Machine Learning. CS Spring 2015 a Bayesian Learning (I) Uncertainty

CS 2740 Knowledge Representation. Lecture 4. Propositional logic. CS 2740 Knowledge Representation. Administration

Introduction to Mobile Robotics Probabilistic Robotics

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018

Building Bayesian Networks. Lecture3: Building BN p.1

CS 5100: Founda.ons of Ar.ficial Intelligence

Probability and Information Theory. Sargur N. Srihari

Uncertainty. 22c:145 Artificial Intelligence. Problem of Logic Agents. Foundations of Probability. Axioms of Probability

Lecture 8. Probabilistic Reasoning CS 486/686 May 25, 2006

Reasoning Under Uncertainty

COMP9414: Artificial Intelligence Reasoning Under Uncertainty

Basic Probability and Decisions

CS 561: Artificial Intelligence

Informatics 2D Reasoning and Agents Semester 2,

CS 343: Artificial Intelligence

Uncertainty Chapter 13. Mausam (Based on slides by UW-AI faculty)

COMP9414/ 9814/ 3411: Artificial Intelligence. 14. Uncertainty. Russell & Norvig, Chapter 13. UNSW c AIMA, 2004, Alan Blair, 2012

CS 188: Artificial Intelligence Spring Announcements

Bayesian Approaches Data Mining Selected Technique

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks. Probability Theory and Probabilistic Reasoning. Emma Rollon and Javier Larrosa Q

Objectives. Probabilistic Reasoning Systems. Outline. Independence. Conditional independence. Conditional independence II.

Probabilistic and Bayesian Analytics

Transcription:

CS 2710 Foundations of AI Lecture 18 Modeling and reasoning with uncertainty Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square KB systems. Medical example. We want to build a KB system for the diagnosis of pneumonia. Problem description: Disease: pneumonia Patient symptoms (findings lab tests: Fever Cough Paleness WBC (white blood cells count Chest pain etc. Representation of a patient case: Statements that hold (are true for the patient. E.g: Fever =True Cough =False WBCcount=High Diagnostic task: we want to decide whether the patient suffers from the pneumonia or not given the symptoms 1

Uncertainty To make diagnostic inference possible we need to represent knowledge (axioms that relate symptoms and diagnosis Paleness Fever Cough WBC count Problem: disease/symptoms relations are not deterministic They are uncertain (or stochastic and vary from patient to patient Uncertainty Two types of uncertainty: Disease Symptoms uncertainty A patient suffering from pneumonia may not have fever all the times may or may not have a cough white blood cell test can be in a normal range. Symptoms Disease uncertainty High fever is typical for many diseases (e.g. bacterial diseases and does not point specifically to pneumonia Fever cough paleness high WBC count combined do not always point to pneumonia 2

Uncertainty Why are relations uncertain? Observability It is impossible to observe all relevant components of the world Observable components behave stochastically even if the underlying world is deterministic Efficiency capacity limits It is often impossible to enumerate and model all components of the world and their relations abstractions can make the relations stochastic Humans can reason with uncertainty!!! Can computer systems do the same? Modeling the uncertainty. Key challenges: How to represent the relations in the presence of uncertainty? How to manipulate such knowledge to make inferences? Humans can reason with uncertainty.? Paleness Fever Cough WBC count 3

Methods for representing uncertainty Extensions of the propositional and first-order logic Use uncertain imprecise statements (relations Example: Propositional logic with certainty factors Very popular in 70-80s in knowledge-based systems (MYCIN Facts (propositional statements are assigned a certainty value reflecting the belief in that the statement is satisfied: CF( True 0.7 Knowledge: typically in terms of modular rules If Then 1. The patient has cough and 2. The patient has a high WBC count and 3. The patient has fever with certainty 0.7 the patient has pneumonia Certainty factors Problem 1: Chaining of multiple inference rules (propagation of uncertainty Solution: Rules incorporate tests on the certainty values ( A in [0.51] ( B in [0.71] C with CF 0.8 Problem 2: Combinations of rules with the same conclusion ( A in [0.51] ( B in [0.71] C with CF 0.8 ( E in [0.81] ( D in [0.91] C with CF 0.9 What is the resulting CF(C? 4

Certainty factors Combination of multiple rules ( A in [0.51] ( B in [0.71] C with CF 0.8 ( E in [0.81] ( D in [0.91] C with CF 0.9 Three possible solutions CF( C max[ 0.9;0.8] 0.9 CF( C 0.9*0.8 0.72 CF( C 0.9 0.8 0.9*0.8 0.98? Problems: Which solution to choose? All three methods break down after a sequence of inference rules Methods for representing uncertainty Probability theory A well defined theory for modeling and reasoning in the presence of uncertainty A natural choice to replace certainty factors Facts (propositional statements Are represented via random variables with two or more values Example: is a random variable values: True and False Each value can be achieved with some probability: True 0.001 WBCcount high 0.005 5

Probability theory Well-defined theory for representing and manipulating statements with uncertainty Axioms of probability: For any two propositions A B. 1. 0 A 1 2. True 1 and False 0 3. A A A Modeling uncertainty with probabilities Probabilistic extension of propositional logic Propositions: statements about the world Represented by the assignment of values to random variables Random variables: Boolean is either True False!! Multi-valued Continuous Random variable Pain is one of Random variable HeartRate Random variable Values { Nopain Mild Moderate Severe} is a value Values in 0; 180 Values 6

Probabilities Unconditional probabilities (prior probabilities 0.001 WBCcount high 0.005 Probability distribution or False 0.999 True 0.001 Defines probabilities for all possible value assignments to a random variable Values are mutually exclusive True 0.001 False 0.999 True False 0.001 0.999 Probability distribution Defines probability for all possible value assignments Example 1: True 0.001 False 0.999 True False True False 1 Example 2: Probabilities sum to 1!!! 0.001 0.999 WBCcount high 0.005 WBCcount WBCcount WBCcount normal 0.993 high 0.005 normal 0.993 WBCcount high 0.002 low 0. 002 7

Joint probability distribution Joint probability distribution (for a set variables Defines probabilities for all possible assignments of values to variables in the set Example 2: Assume variables: (2 values WBCcount (3 values Pain (4 values pneumonia WBCcount Pain is represented by 234 array Joint probability distribution Joint probability distribution (for a set variables Defines probabilities for all possible assignments of values to variables in the set Example: variables and WBCcount pneumonia WBCcount is represented by 23 matrix True False WBCcount high normal low 0.0008 0.0042 0.0001 0.9929 0.0001 0.0019 8

Joint probabilities Marginalization reduces the dimension of the joint distribution Sums variables out pneumonia WBCcount True False 23 matrix WBCcount high normal low 0.0008 0.0001 0.0001 0.0042 0.9929 0.0019 0.005 0. 993 0. 002 0.001 0.999 WBCcount Marginalization (here summing of columns or rows Marginalization Marginalization reduces the dimension of the joint distribution X 1 X 2 X n1 X1 X 2 X n1 X n { } X n We can continue doing this X 2 X n1 { X 1 X n X } 1 X 2 X n1 X n What is the maximal joint probability distribution? Full joint probability 9

Full joint distribution The joint distribution for all variables in the problem It defines the complete probability model for the problem Example: pneumonia diagnosis Variables: Fever Paleness WBCcount Cough Full joint defines the probability for all possible assignments of values to these variables T WBCcount High Fever T Cough T Paleness T T WBCcount High Fever T Cough T Paleness F T WBCcount High Fever T Cough F Paleness T etc How many probabilities are there? Full joint distribution the joint distribution for all variables in the problem It defines the complete probability model for the problem Example: pneumonia diagnosis Variables: Fever Paleness WBCcount Cough Full joint defines the probability for all possible assignments of values to these variables T WBCcount High Fever T Cough T Paleness T T WBCcount High Fever T Cough T Paleness F T WBCcount High Fever T Cough F Paleness T etc How many probabilities are there? Exponential in the number of variables 10

Full joint distribution Any joint probability for a subset of variables can be obtained via marginalization WBCcount Fever c p{ T F} WBCcount Fever Cough c Paleness Is it possible to recover the full joint from the joint probabilities over a subset of variables? p Joint probabilities Is it possible to recover the full joint from the joint probabilities over a subset of variables? pneumonia WBCcount True False 23 matrix WBCcount high normal low?????? 0.005 0. 993 0. 002 0.001 0.999 WBCcount 11

Joint probabilities and independence Is it possible to recover the full joint from the joint probabilities over a subset of variables? Only if the variables are independent!!! pneumonia WBCcount True False WBCcount 23 matrix WBCcount high normal low?????? 0.005 0. 993 0. 002 0.001 0.999 Conditional probabilities Conditional probability distribution. A A s.t. 0 Product rule. Join probability can be expressed in terms of conditional probabilities A A Chain rule. Any joint probability can be expressed as a product of conditionals X1 X 2 X n X n X1 X n1 X1 X n1 X n X1 X n1 X n1 X1 X n2 X1 X n2 n X i i X X 1 1 i1 12

Conditional probabilities Conditional probability Is defined in terms of the joint probability: A A s.t. 0 Example: pneumonia true WBCcount high pneumonia true WBCcount high WBCcount high pneumonia false WBCcount high pneumonia false WBCcount high WBCcount high Conditional probabilities Conditional probability distribution Defines probabilities for all possible assignments given a fixed assignment to some other variable values true WBCcount high WBCcount 3 element vector of 2 elements WBCcount high normal low True False true WBCcount high 0.08 0.0001 0.0001 0.92 0.9999 0.9999 1.0 1. 0 1. 0 false WBCcount high 13

Conditional probability. Bayes rule A A A B A A Bayes rule: B A A A When is it useful? When we are interested in computing the diagnostic query from the causal probability effect cause cause cause effect effect Reason: It is often easier to assess causal probability E.g. Probability of pneumonia causing fever vs. probability of pneumonia given fever Various inference tasks: Probabilistic inference Diagnostic task. (from effect to cause Fever T Prediction task. (from cause to effect Fever T Other probabilistic queries (queries on joint distributions. Fever Fever ChestPain 14

15 Inference Any query can be computed from the full joint distribution!!! Joint over a subset of variables is obtained through marginalization Conditional probability over set of variables given other variables values is obtained through marginalization and definition of conditionals ( ( i j i d j D c C b B a A P c C a A P ( ( i j j i i i d D c C b B a A P d D c C b B a A P ( ( ( c C a A P d D c C a A P c C a A d D P Modeling uncertainty with probabilities Defining the full joint distribution makes it possible to represent and reason with uncertainty in a uniform way We are able to handle an arbitrary inference problem Problems: Space complexity. To store a full joint distribution we need to remember numbers. n number of random variables d number of values Inference (time complexity. To compute some queries requires. steps. Acquisition problem. Who is going to define all of the probability entries? O(d n O(d n

Medical diagnosis example Space complexity. (2 values: TF Fever (2: TF Cough (2: TF WBCcount (3: high normal low paleness (2: TF Number of assignments: 2*2*2*3*2=48 We need to define at least 47 probabilities. Time complexity. Assume we need to compute the marginal of =T from the full joint T Fever i Cough it F jt F kh n l ut F Sum over: 2*2*3*2=24 combinations j WBCcount k Pale u Modeling uncertainty with probabilities Knowledge based system era (70s early 80 s Extensional non-probabilistic models Solve the space time and acquisition bottlenecks in probability-based models froze the development and advancement of KB systems and contributed to the slow-down of AI in 80s in general Breakthrough (late 80s beginning of 90s Bayesian belief networks Give solutions to the space acquisition bottlenecks Partial solutions for time complexities Bayesian belief network 16

Bayesian belief networks (BBNs Bayesian belief networks. Represent the full joint distribution over the variables more compactly with a smaller number of parameters. Take advantage of conditional and marginal independences among random variables A and B are independent A A A and B are conditionally independent given C A B C A C B C A C A C 17