Module 10. Reasoning with Uncertainty - Probabilistic reasoning. Version 2 CSE IIT,Kharagpur

Similar documents
UNCERTAIN KNOWLEDGE AND REASONING

1) partial observability (road state, other drivers' plans, etc.) 4) immense complexity of modelling and predicting tra_c

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Motivation. Bayesian Networks in Epistemology and Philosophy of Science Lecture. Overview. Organizational Issues

Lecture 5: Bayesian Network

Introduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah

Bayesian Networks and Decision Graphs

Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence

Introduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah

Introduction to Artificial Intelligence. Unit # 11

Semirings for Breakfast

TDT70: Uncertainty in Artificial Intelligence. Chapter 1 and 2

Probabilistic Models

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber

STATISTICAL METHODS IN AI/ML Vibhav Gogate The University of Texas at Dallas. Propositional Logic and Probability Theory: Review

COS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference

Lecture 1. ABC of Probability

Uncertainty and Bayesian Networks

Support for Multiple Cause Diagnosis with Bayesian Networks

Econ 325: Introduction to Empirical Economics

Bayesian Reasoning. Adapted from slides by Tim Finin and Marie desjardins.

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Modeling and reasoning with uncertainty

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

CS 5522: Artificial Intelligence II

Probabilistic Reasoning; Network-based reasoning

COMP538: Introduction to Bayesian Networks

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018

Reasoning with Inconsistent and Uncertain Ontologies

Probabilistic Reasoning; Network-based reasoning

CS 343: Artificial Intelligence

Human-level concept learning through probabilistic program induction

CS 188: Artificial Intelligence Fall 2008

Uncertain Knowledge and Bayes Rule. George Konidaris

Exact Inference Algorithms Bucket-elimination

CSE 473: Artificial Intelligence Autumn 2011

Logic: Propositional Logic Truth Tables

CS 188: Artificial Intelligence Fall 2009

Computer Science CPSC 322. Lecture 23 Planning Under Uncertainty and Decision Networks

Pengju XJTU 2016

Pengju

Exact Inference Algorithms Bucket-elimination

Bayesian Networks for Causal Analysis

Conservative Inference Rule for Uncertain Reasoning under Incompleteness

Basic Probabilistic Reasoning SEG

Our learning goals for Bayesian Nets

CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-II

Bayesian Networks. Probability Theory and Probabilistic Reasoning. Emma Rollon and Javier Larrosa Q

[read Chapter 2] [suggested exercises 2.2, 2.3, 2.4, 2.6] General-to-specific ordering over hypotheses

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Modal Logics. Most applications of modal logic require a refined version of basic modal logic.

Probabilistic Models. Models describe how (a portion of) the world works

Fuzzy Systems. Possibility Theory.

CS 188: Artificial Intelligence Spring Announcements

02 Propositional Logic

Conditional Independence

AI Programming CS S-09 Knowledge Representation

Lecture 16 : Independence, Covariance and Correlation of Discrete Random Variables

Building Bayesian Networks. Lecture3: Building BN p.1

Quantifying Uncertainty & Probabilistic Reasoning. Abdulla AlKhenji Khaled AlEmadi Mohammed AlAnsari

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012

What Causality Is (stats for mathematicians)

Uncertainty. Chapter 13

Logic: The Big Picture

Introduction to Bayesian Networks

Introduction to Machine Learning

Probability Theory for Machine Learning. Chris Cremer September 2015

Generative Techniques: Bayes Rule and the Axioms of Probability

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

Review of Basic Probability

Uncertainty. Logic and Uncertainty. Russell & Norvig. Readings: Chapter 13. One problem with logical-agent approaches: C:145 Artificial

Reasoning Under Uncertainty

Axioms of Probability? Notation. Bayesian Networks. Bayesian Networks. Today we ll introduce Bayesian Networks.

A Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004

Probabilistic Reasoning

Resolution or modus ponens are exact there is no possibility of mistake if the rules are followed exactly.

CPSC340. Probability. Nando de Freitas September, 2012 University of British Columbia

Probability Basics. Robot Image Credit: Viktoriya Sukhanova 123RF.com

Directed Graphical Models or Bayesian Networks

CS 2740 Knowledge Representation. Lecture 4. Propositional logic. CS 2740 Knowledge Representation. Administration

Chapter 4: Classical Propositional Semantics

Uncertainty. Outline

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

An Introduction to Bayesian Networks: Representation and Approximate Inference

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation

Reasoning Under Uncertainty: Introduction to Probability

CS4705. Probability Review and Naïve Bayes. Slides from Dragomir Radev

Introduction to Bayesian Networks. Probabilistic Models, Spring 2009 Petri Myllymäki, University of Helsinki 1

Discrete Probability and State Estimation

Framing human inference by coherence based probability logic

Artificial Intelligence Programming Probability

Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule

Introduction to Mobile Robotics Probabilistic Robotics

Probabilistic Reasoning; Network-based reasoning

Probabilistic Robotics

Bayesian Networks. Marcello Cirillo PhD Student 11 th of May, 2007

Bayesian Networks Representation

Introduction to Causal Calculus

Transcription:

Module 10 Reasoning with Uncertainty - Probabilistic reasoning

10.1 Instructional Objective The students should understand the role of uncertainty in knowledge representation Students should learn the use of probability theory to represent uncertainty Students should understand the basic of probability theory, including o Probability distributions o Joint probability o Marginal probability o Conditional probability o Independence o Conditional independence Should learn inference mechanisms in probability theory including o Bayes rule o Product rule Should be able to convert natural language statements into probabilistic statements and apply inference rules Students should understand Bayesian networks as a data structure to represent conditional independence Should understand the syntax and semantics of Bayes net Should understand inferencing mechanisms in Bayes net Should understand efficient inferencing techniques like variable ordering Should understand the concept of d-separation Should understand inference mechanism for the special case of polytrees Students should have idea about approximate inference techniques in Bayesian networks At the end of this lesson the student should be able to do the following: Represent a problem in terms of probabilistic statemenst Apply Bayes rule and product rule for inferencing Represent a problem using Bayes net Perform probabilistic inferencing using Bayes net.

Lesson 26 Reasoning with Uncertain information

10. 2 Probabilistic Reasoning Using logic to represent and reason we can represent knowledge about the world with facts and rules, like the following ones: bird(tweety). fly(x) :- bird(x). We can also use a theorem-prover to reason about the world and deduct new facts about the world, for e.g.,?- fly(tweety). Yes However, this often does not work outside of toy domains - non-tautologous certain rules are hard to find. A way to handle knowledge representation in real problems is to extend logic by using certainty factors. In other words, replace IF condition THEN fact with IF condition with certainty x THEN fact with certainty f(x) Unfortunately cannot really adapt logical inference to probabilistic inference, since the latter is not context-free. Replacing rules with conditional probabilities makes inferencing simpler. Replace smoking -> lung cancer or lotsofconditions, smoking -> lung cancer with P(lung cancer smoking) = 0.6 Uncertainty is represented explicitly and quantitatively within probability theory, a formalism that has been developed over centuries. A probabilistic model describes the world in terms of a set S of possible states - the sample space. We don t know the true state of the world, so we (somehow) come up with a probability distribution over S which gives the probability of any state being the true one. The world usually described by a set of variables or attributes. Consider the probabilistic model of a fictitious medical expert system. The world is described by 8 binary valued variables:

Visit to Asia? A Tuberculosis? T Either tub. or lung cancer? E Lung cancer? L Smoking? S Bronchitis? B Dyspnoea? D Positive X-ray? X We have 2 8 = 256 possible states or configurations and so 256 probabilities to find. 10.3 Review of Probability Theory The primitives in probabilistic reasoning are random variables. Just like primitives in Propositional Logic are propositions. A random variable is not in fact a variable, but a function from a sample space S to another space, often the real numbers. For example, let the random variable Sum (representing outcome of two die throws) be defined thus: Sum(die1, die2) = die1 +die2 Each random variable has an associated probability distribution determined by the underlying distribution on the sample space Continuing our example : P(Sum = 2) = 1/36, P(Sum = 3) = 2/36,..., P(Sum = 12) = 1/36 Consdier the probabilistic model of the fictitious medical expert system mentioned before. The sample space is described by 8 binary valued variables. Visit to Asia? A Tuberculosis? T Either tub. or lung cancer? E Lung cancer? L Smoking? S Bronchitis? B Dyspnoea? D Positive X-ray? X There are 2 8 = 256 events in the sample space. Each event is determined by a joint instantiation of all of the variables. S = {(A = f, T = f,e = f,l = f, S = f,b = f,d = f,x = f), (A = f, T = f,e = f,l = f, S = f,b = f,d = f,x = t),... (A = t, T = t,e = t,l = t, S = t,b = t,d = t,x = t)}

Since S is defined in terms of joint instantations, any distribution defined on it is called a joint distribution. ll underlying distributions will be joint distributions in this module. The variables {A,T,E, L,S,B,D,X} are in fact random variables, which project values. L(A = f, T = f,e = f,l = f, S = f,b = f,d = f,x = f) = f L(A = f, T = f,e = f,l = f, S = f,b = f,d = f,x = t) = f L(A = t, T = t,e = t,l = t, S = t,b = t,d = t,x = t) = t Each of the random variables {A,T,E,L,S,B,D,X} has its own distribution, determined by the underlying joint distribution. This is known as the margin distribution. For example, the distribution for L is denoted P(L), and this distribution is defined by the two probabilities P(L = f) and P(L = t). For example, P(L = f) = P(A = f, T = f,e = f,l = f, S = f,b = f,d = f,x = f) + P(A = f, T = f,e = f,l = f, S = f,b = f,d = f,x = t) + P(A = f, T = f,e = f,l = f, S = f,b = f,d = t,x = f)... P(A = t, T = t,e = t,l = f, S = t,b = t,d = t,x = t) P(L) is an example of a marginal distribution. Here s a joint distribution over two binary value variables A and B We get the marginal distribution over B by simply adding up the different possible values of A for any value of B (and put the result in the margin ). In general, given a joint distribution over a set of variables, we can get the marginal distribution over a subset by simply summing out those variables not in the subset.

In the medical expert system case, we can get the marginal distribution over, say, A,D by simply summing out the other variables: However, computing marginals is not an easy task always. For example, P(A = t,d = f) = P(A = t, T = f,e = f,l = f, S = f,b = f,d = f,x = f) + P(A = t, T = f,e = f,l = f, S = f,b = f,d = f,x = t) + P(A = t, T = f,e = f,l = f, S = f,b = t,d = f,x = f) + P(A = t, T = f,e = f,l = f, S = f,b = t,d = f,x = t)... P(A = t, T = t,e = t,l = t, S = t,b = t,d = f,x = t) This has 64 summands! Each of whose value needs to be estimated from empirical data. For the estimates to be of good quality, each of the instances that appear in the summands should appear sufficiently large number of times in the empirical data. Often such a large amount of data is not available. However, computation can be simplified for certain special but common conditions. This is the condition of independence of variables. Two random variables A and B are independent iff P(A,B) = P(A)P(B) i.e. can get the joint from the marginals This is quite a strong statement: It means for any value x of A and any value y of B P(A = x,b = y) = P(A = x)p(b = y) Note that the independence of two random variables is a property of a the underlying probability distribution. We can have Conditional probability is defined as: It means for any value x of A and any value y of B

If A and B are independent then Conditional probabilities can represent causal relationships in both directions. From cause to (probable) effects From effect to (probable) cause