Reasoning Under Uncertainty: More on BNets structure and construction

Similar documents
Reasoning Under Uncertainty: More on BNets structure and construction

Intelligent Systems (AI-2)

COS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference

Reasoning Under Uncertainty: Belief Network Inference

Informatics 2D Reasoning and Agents Semester 2,

Probability and Time: Hidden Markov Models (HMMs)

Directed Graphical Models

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

CS 5522: Artificial Intelligence II

Intelligent Systems (AI-2)

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

Introduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013

Introduction to Artificial Intelligence (AI)

Probabilistic Reasoning. (Mostly using Bayesian Networks)

CS 188: Artificial Intelligence Fall 2009

Bayesian Networks. Semantics of Bayes Nets. Example (Binary valued Variables) CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-III

CS Lecture 3. More Bayesian Networks

Intelligent Systems (AI-2)

Lecture 8. Probabilistic Reasoning CS 486/686 May 25, 2006

Rapid Introduction to Machine Learning/ Deep Learning

Product rule. Chain rule

Bayesian Inference. Definitions from Probability: Naive Bayes Classifiers: Advantages and Disadvantages of Naive Bayes Classifiers:

Reasoning under Uncertainty: Intro to Probability

Probabilistic Graphical Networks: Definitions and Basic Results

CS 188: Artificial Intelligence Spring Announcements

Bayes Nets: Independence

Probabilistic Classification

CS 188: Artificial Intelligence Fall 2008

Intelligent Systems (AI-2)

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation

CS 484 Data Mining. Classification 7. Some slides are from Professor Padhraic Smyth at UC Irvine

Directed and Undirected Graphical Models

Artificial Intelligence Bayesian Networks

Directed Graphical Models or Bayesian Networks

Introduction to Artificial Intelligence (AI)

Bayesian belief networks. Inference.

Axioms of Probability? Notation. Bayesian Networks. Bayesian Networks. Today we ll introduce Bayesian Networks.

Reasoning Under Uncertainty: Bayesian networks intro

Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence

Review: Bayesian learning and inference

Directed and Undirected Graphical Models

CS 188: Artificial Intelligence Spring Announcements

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

PROBABILISTIC REASONING SYSTEMS

Artificial Intelligence Bayes Nets: Independence

Reasoning under Uncertainty: Intro to Probability

Bayesian Approaches Data Mining Selected Technique

Bayesian Network. Outline. Bayesian Network. Syntax Semantics Exact inference by enumeration Exact inference by variable elimination

Uncertainty and Bayesian Networks

Probabilistic Models

CS 5522: Artificial Intelligence II

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Which coin will I use? Which coin will I use? Logistics. Statistical Learning. Topics. 573 Topics. Coin Flip. Coin Flip

Bayesian Networks BY: MOHAMAD ALSABBAGH

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012

Bayesian belief networks

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

Intro to AI: Lecture 8. Volker Sorge. Introduction. A Bayesian Network. Inference in. Bayesian Networks. Bayesian Networks.

CS 343: Artificial Intelligence

Bayesian Networks Representation

Bayesian Networks. Axioms of Probability Theory. Conditional Probability. Inference by Enumeration. Inference by Enumeration CSE 473

Building Bayesian Networks. Lecture3: Building BN p.1

Introduction to Bayesian Learning

Bayesian Networks. Machine Learning, Fall Slides based on material from the Russell and Norvig AI Book, Ch. 14

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

Quantifying uncertainty & Bayesian networks

Tópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863

Based on slides by Richard Zemel

CS839: Probabilistic Graphical Models. Lecture 2: Directed Graphical Models. Theo Rekatsinas

Representation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2

Introduction to Artificial Intelligence. Unit # 11

CS6220: DATA MINING TECHNIQUES

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Bayes Networks. CS540 Bryan R Gibson University of Wisconsin-Madison. Slides adapted from those used by Prof. Jerry Zhu, CS540-1

Announcements. CS 188: Artificial Intelligence Spring Bayes Net Semantics. Probabilities in BNs. All Conditional Independences

CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-II

Stephen Scott.

Introduction to Probabilistic Reasoning. Image credit: NASA. Assignment

CSE 473: Artificial Intelligence Probability Review à Markov Models. Outline

An AI-ish view of Probability, Conditional Probability & Bayes Theorem

10/18/2017. An AI-ish view of Probability, Conditional Probability & Bayes Theorem. Making decisions under uncertainty.

Reasoning Under Uncertainty: Conditional Probability

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

Reasoning Under Uncertainty: Conditioning, Bayes Rule & the Chain Rule

Bayes Nets. CS 188: Artificial Intelligence Fall Example: Alarm Network. Bayes Net Semantics. Building the (Entire) Joint. Size of a Bayes Net

CSE 473: Artificial Intelligence Autumn 2011

14 PROBABILISTIC REASONING

Bayesian Networks. Alan Ri2er

Learning in Bayesian Networks

Bayesian Belief Network

Recap: Bayes Nets. CS 473: Artificial Intelligence Bayes Nets: Independence. Conditional Independence. Bayes Nets. Independence in a BN

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018

Probabilistic Graphical Models

CSCE 478/878 Lecture 6: Bayesian Learning and Graphical Models. Stephen Scott. Introduction. Outline. Bayes Theorem. Formulas

CS 188: Artificial Intelligence Spring Today

Probabilistic Models. Models describe how (a portion of) the world works

Bayesian networks (1) Lirong Xia

Preliminaries Bayesian Networks Graphoid Axioms d-separation Wrap-up. Bayesian Networks. Brandon Malone

Outline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

Artificial Intelligence: Reasoning Under Uncertainty/Bayes Nets

Transcription:

Reasoning Under Uncertainty: More on BNets structure and construction Jim Little Nov 10 2014 (Textbook 6.3) Slide 1

Belief networks Recap By considering causal dependencies, we order variables in the joint. Apply? and simplify chain rule Build a directed acyclic graph (DAG) in which the parents of each var X are those vars on which X directly depends. By construction, a var is independent from its non-descendants given its parents. Slide 2

Belief Networks: open issues Independencies: Does a BNet encode more independencies than the ones specified by construction? Compactness: For n vars, k parents, we reduce the number of probabilities from to In some domains we need to do better than that! Still too many and often there are no data/experts for accurate assessment Solution: Make stronger (approximate) independence assumptions Slide 3

Lecture Overview Implied Conditional Independence relations in a Bnet Compactness: Making stronger Independence assumptions Representation of Compact Conditional Distributions Network structure( Naïve Bayesian Classifier) Slide 4

Bnets: Entailed (in)dependencies Indep(Report, Fire,{Alarm})? Indep(Leaving, SeeSmoke,{Fire})? Slide 5

Conditional Independencies Or, blocking paths for probability propagation. Three ways in which a path between X to Y can be blocked, (1 and 2 given evidence E ) 1 Y E X 2 3 Note that, in 3, X and Y become dependent as soon as I get evidence on or on any of its descendants Slide 6

Or.Conditional Dependencies 1 Y X 2 3 E In 1, 2, 3 X Y are dependent Slide 7

In/Dependencies in a Bnet : Example 1 1 Y E X 2 3 Is A conditionally independent of I given F? Slide 8

In/Dependencies in a Bnet : Example 2 1 Y E X 2 3 Slide 9

In/Dependencies in a Bnet : Example 3 1 Y E X 2 3 Slide 10

Lecture Overview Implied Conditional Independence relations in a Bnet Compactness: Making stronger Independence assumptions Representation of Compact Conditional Distributions Network structure( Naïve Bayesian Classifier) Slide 11

More on Construction and Compactness: Compact Conditional Distributions Once we have established the topology of a Bnet, we still need to specify the conditional probabilities How? From Data From Experts To facilitate acquisition, we aim for compact representations for which data/experts can provide accurate assessments Slide 12

More on Construction and Compactness: Compact Conditional Distributions From JointPD to But still, CPT grows exponentially with number of parents In realistic model of internal medicine with 448 nodes and 906 links 133,931,430 values are required! And often there are no data/experts for accurate assessment Slide 13

Effect with multiple non-interacting causes Malaria Flu Cold What do we need to specify? Malaria Flu Cold P(Fever=T..) P(Fever=F..) Fever What do you think data/experts could easily tell you? T T T T T F T F T T F F F T T F T F F F T F F F More difficult to get info to assess more complex conditioning. Slide 14

Solution: Noisy-OR Distributions Models multiple non interacting causes Logic OR with a probabilistic twist. Logic OR Conditional Prob. Table. Malaria Flu Cold P(Fever=T..) P(Fever=F..) T T T T T F T F T T F F F T T F T F F F T F F F Slide 15

Solution: Noisy-OR Distributions The Noisy-OR model allows for uncertainty in the ability of each cause to generate the effect (e.g.. one may have a cold without a fever) Malaria Flu Cold P(Fever=T..) P(Fever=F..) T T T T T F T F T T F F F T T F T F F F T F F F Two assumptions 1. All possible causes are listed 2. For each of the causes, whatever inhibits it to generate the target effect is independent from the inhibitors of the other causes Slide 16

Noisy-OR: Derivations C 1 C k For each of the causes, whatever inhibits it from generating the target effect is independent from the inhibitors of the other causes Effect Independent Probability of failure q i for each cause alone: P(Effect=F C i = T, and no other causes) = q i P(Effect=F C 1 = T,.. C j = T, C j+1 = F,., C k = F)= A. B. C. D. None of these Slide 17

Noisy-OR: Derivations C 1 C k Effect For each of the causes, whatever inhibits it to generate the target effect is independent from the inhibitors of the other causes Independent Probability of failure q i for each cause alone: P(Effect=F C i = T, and no other causes) = q i P(Effect=F C 1 = T,.. C j = T, C j+1 = F,., C k = F)= P(Effect=T C 1 = T,.. C j = T, C j+1 = F,., C k = F) = 1- Slide 18

Noisy-OR: Example P(Fever=F Cold=T, Flu=F, Malaria=F) = 0.6 P(Fever=F Cold=F, Flu=T, Malaria=F) = 0.2 P(Fever=F Cold=F, Flu=F, Malaria=T) = 0.1 Model of internal medicine 133,931,430 8,254 With noisy ors P(Effect=F C 1 = T,.. C j = T, C j+1 = F,., C k = F)= j i=1 q i Malaria Flu Cold P(Fever=T..) P(Fever=F..) T T T 0.1 x 0.2 x 0.6 = 0.012 T T F 0.2 x 0.1 = 0.02 T F T 0.6 x 0.1=0.06 T F F 0.9 0.1 F T T 0.2 x 0.6 = 0.12 F T F 0.8 0.2 F F T 0.4 0.6 F F F 1.0 Number of probabilities needed linear in k Slide 19 3 here

Lecture Overview Implied Conditional Independence relations in a Bnet Compactness: Making stronger Independence assumptions Representation of Compact Conditional Distributions Network structure ( Naïve Bayesian Classifier) Slide 20

Naïve Bayesian Classifier A very simple and successful form of Bnet that can classify entities in a set of classes C, given a set of attributes Example: Determine whether an email is spam (only two classes spam=t and spam=f) Useful attributes of an email? Assumptions The value of each attribute depends on the classification (Naïve) The attributes are independent of each other given the classification Slide 21 P( bank account, spam=t) = P( bank spam=t)

A. What is the structure? Assumptions The value of each attribute depends on the classification (Naïve) The attributes are independent of each other given the classification Email Spam free money ubc midterm B. Email Spam free money ubc midterm free C. money ubc midterm Email Spam Slide 22

Naïve Bayesian Classifier for Email Spam Assumptions The value of each attribute depends on the classification (Naïve) The attributes are independent of each other given the classification Email Spam words free money ubc midterm Number of parameters? Easy to acquire? If you have a large collection of emails for which Slide 23 you know if they are spam or not

NB Classifier for Email Spam: Usage Most likely class given set of observations Is a given Email E spam? free money for you now Email Spam free money ubc midterm Email is a spam if. Slide 24

For another example of naïve Bayesian See textbook ex. 6.16 Classifier help system to determine what help page a user is interested in based on the keywords they give in a query to a help system. Slide 25

You can: Learning Goals for today s class Given a Belief Net, determine whether one variable is conditionally independent of another variable, given a set of observations. Define and use Noisy-OR distributions. Explain assumptions and benefit. Implement and use a naïve Bayesian classifier. Explain assumptions and benefit. Slide 26

Next Class Bayesian Networks Inference: Variable Elimination Slide 27