Events A and B are independent P(A) = P(A B) = P(A B) / P(B)

Similar documents
Uncertainty and Belief Networks. Introduction to Artificial Intelligence CS 151 Lecture 1 continued Ok, Lecture 2!

Probabilistic Models

CS 188: Artificial Intelligence Fall 2008

Probabilistic Models. Models describe how (a portion of) the world works

CS 188: Artificial Intelligence Spring Announcements

CS 484 Data Mining. Classification 7. Some slides are from Professor Padhraic Smyth at UC Irvine

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

CS 5522: Artificial Intelligence II

Introduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013

CS 5522: Artificial Intelligence II

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

CMPSCI 240: Reasoning about Uncertainty

CS 343: Artificial Intelligence

This lecture. Reading. Conditional Independence Bayesian (Belief) Networks: Syntax and semantics. Chapter CS151, Spring 2004

CS 188: Artificial Intelligence Fall 2009

CSE 473: Artificial Intelligence Autumn 2011

Bayes Nets: Independence

Bayesian Networks. Semantics of Bayes Nets. Example (Binary valued Variables) CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-III

Bayesian networks (1) Lirong Xia

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Directed Graphical Models

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

CS839: Probabilistic Graphical Models. Lecture 2: Directed Graphical Models. Theo Rekatsinas

Directed Graphical Models or Bayesian Networks

Objectives. Probabilistic Reasoning Systems. Outline. Independence. Conditional independence. Conditional independence II.

Lecture 8. Probabilistic Reasoning CS 486/686 May 25, 2006

Probabilistic Reasoning Systems

Quantifying Uncertainty & Probabilistic Reasoning. Abdulla AlKhenji Khaled AlEmadi Mohammed AlAnsari

13.4 INDEPENDENCE. 494 Chapter 13. Quantifying Uncertainty

PROBABILISTIC REASONING SYSTEMS

Bayesian Belief Network

Quantifying uncertainty & Bayesian networks

Review: Bayesian learning and inference

Artificial Intelligence Bayes Nets: Independence

Axioms of Probability? Notation. Bayesian Networks. Bayesian Networks. Today we ll introduce Bayesian Networks.

Probabilistic Classification

Cartesian-product sample spaces and independence

Graphical Models - Part I

Bayesian Networks. Material used

Uncertainty. Chapter 13

Bayesian Networks BY: MOHAMAD ALSABBAGH

In today s lecture. Conditional probability and independence. COSC343: Artificial Intelligence. Curse of dimensionality.

Probabilistic representation and reasoning

Bayesian networks. Chapter 14, Sections 1 4

Rapid Introduction to Machine Learning/ Deep Learning

Informatics 2D Reasoning and Agents Semester 2,

Reasoning under Uncertainty: Intro to Probability

UNCERTAINTY. In which we see what an agent should do when not all is crystal-clear.

Bayesian networks. Chapter Chapter

Y. Xiang, Inference with Uncertain Knowledge 1

Uncertainty. Introduction to Artificial Intelligence CS 151 Lecture 2 April 1, CS151, Spring 2004

Bayesian Networks. Machine Learning, Fall Slides based on material from the Russell and Norvig AI Book, Ch. 14

ARTIFICIAL INTELLIGENCE. Uncertainty: probabilistic reasoning

Conditional Independence

Fusion in simple models

Reasoning Under Uncertainty: Bayesian networks intro

Uncertainty. Variables. assigns to each sentence numerical degree of belief between 0 and 1. uncertainty

Bayesian Networks aka belief networks, probabilistic networks. Bayesian Networks aka belief networks, probabilistic networks. An Example Bayes Net

Reasoning Under Uncertainty

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.

Stochastic Methods. 5.0 Introduction 5.1 The Elements of Counting 5.2 Elements of Probability Theory

CS 188: Artificial Intelligence Spring Announcements

Bayesian networks. Independence. Bayesian networks. Markov conditions Inference. by enumeration rejection sampling Gibbs sampler

Implementing Machine Reasoning using Bayesian Network in Big Data Analytics

CS 188: Artificial Intelligence. Our Status in CS188

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

Reasoning with Bayesian Networks

CSE 473: Ar+ficial Intelligence. Hidden Markov Models. Bayes Nets. Two random variable at each +me step Hidden state, X i Observa+on, E i

Bayesian Network. Outline. Bayesian Network. Syntax Semantics Exact inference by enumeration Exact inference by variable elimination

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

2 : Directed GMs: Bayesian Networks

Today. Why do we need Bayesian networks (Bayes nets)? Factoring the joint probability. Conditional independence

Bayesian networks: Modeling

Lecture 4 October 18th

Probabilistic Reasoning. Kee-Eung Kim KAIST Computer Science

Uncertainty and Bayesian Networks

Lecture 8: Bayesian Networks

Uncertainty. Chapter 13

Directed and Undirected Graphical Models

School of EECS Washington State University. Artificial Intelligence

CS Belief networks. Chapter

Foundations of Artificial Intelligence

Computer Science CPSC 322. Lecture 18 Marginalization, Conditioning

Reasoning Under Uncertainty

Directed Graphical Models

Outline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

Brief Intro. to Bayesian Networks. Extracted and modified from four lectures in Intro to AI, Spring 2008 Paul S. Rosenbloom

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models

Pengju

Exercises, II part Exercises, II part

CS 380: ARTIFICIAL INTELLIGENCE UNCERTAINTY. Santiago Ontañón

Probabilistic Graphical Models. Rudolf Kruse, Alexander Dockhorn Bayesian Networks 153

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

CS6220: DATA MINING TECHNIQUES

Bayesian Networks Representation

Transcription:

Events A and B are independent A B U P(A) = P(A B) = P(A B) / P(B) 1

Alternative Characterization of Independence a) P(A B) = P(A) b) P(A B) = P(A) P(B) Recall P(A B) = P(A B) / P(B) (if P(B) 0) So P(A B) / P(B) = P(A) Multiply both sides by P(B) b) Is more obviously symmetric & avoids P(B) 0 a) Gives clearer insight into information about B being irrelevant to estimating A 2

Indepenences are Redundancies in the Joint Random variables A and B are Independent iff B) = P(A) B provides no information about A P(A Suppose Weather distinctions: m Fun levels: n Times of the day: k How many numbers If random variables interact? (m n k) - 1 If they are independent? (m-1) + (n-1) + (k-1) 3

Implications for AI Representations Joint contains just as many numbers Some are functions of others AI models should allow all and only the distinctions necessary for the task Fewer distinctions fewer numbers more efficient processing faster learning fewer inconsistencies smaller data structures 4

Diseases and Symptoms Are they independent? We hope not! What could be independent? Where do the symptoms come from? 5

Conditional Independence Model the disease influencing the symptoms But perhaps no symptom interactions given the disease Conditional independence: P(A B,C) = P(A C) A is conditionally independent of B given C 6

More Conditional Independence Suppose A is conditionally independent of B given C Is B necessarily conditionally independent of A given C? P(A B,C)=P(A C)? P(B A,C)=P(B C) P(A B,C) = P(A C) P(A,B,C) P(A,C) = P(B,C) P(C) P(A,B,C) P(A,C) P(B,C) = P(C) P(B A,C) = P(B C) YES, they are equivalent 7

More Conditional Independence A is conditionally independent of B given C C can still depend on both A and B P(C A) P(C) and P(C B) P(C) A and B are not necessarily independent: P(B A) P(B) and P(A B) P(A) But, Given C, A and B do not influence each other Knowledge of C separates A and B Suppose we do NOT know C, then A and B MAY influence each other (think about this one until it s intuitive what does influence mean here?) 8

Suppose that Symptoms are Conditionally Independent given the Disease We know John has a cold Congestion, a sore throat, headache, rash are more likely but not necessary P(congestion cold) >> P(congestion) likewise sore throat We discover he does in fact have a sore throat Given his cold, is he now more or less likely to also have congestion? Not much: P(congestion sorethroat, cold) ~ P(congestion cold) We may choose not to model this and other weak interactions 9

Bayesian Networks (belief networks, ) Graphical model Probability + Graph theory Nodes discrete random variables Directed Arcs direct influence Usually causality (class, text, ) Configuration of parents establishes contexts for a node Different distribution for each context Conditional Probability Tables (CPTs) No directed cycles (DAG) General graphical models Markov Networks (Markov random fields) Conditional random fields Dynamic Bayesian nets others 10

Dentist Example 3 Boolean Random Variables: C Patient has a cavity A Patient reports a toothache B Dentist s probe catches on tooth C directly influences A and B A C B A and B are conditionally independent given C CPT at each node specifies a probability distribution for each context (configuration of parents) How many numbers do we need for the full joint? How many for the Bayesian net? 11

How Many Numbers Joint: 2 3 1 = 7 (why?) Bayes Net: A Distribution(s) at C P(cavity), P( cavity) Need one number Distribution(s) at A P(ache cavity), P( ache cavity) P(ache cavity), P( ache cavity) Distributions at B, like A 1 + 2 + 2 = 5 (a savings of 2!) C B 12

How Many Numbers? Add binary random variables for Weather W Weekday / Weekend D How many numbers for the joint? 31 How many numbers for the BN? 7 What does the BN look like? W A C B D 13

How Many Numbers? Suppose A,B,C,D,W are trinary How many numbers for the joint? 3 5 1 = 242 How many numbers for the BN? 2+2+2+6+6 = 18 Suppose A is binary, B is trinary, C takes on 4 values D takes on 5, W takes on 6? W 5, D 4, C 3, A 4, B 8 = 24 W A C B D 14

Notation W, C, D, A, B are random variables C (binary presence of cavity) ranges over values C s values might be Yes or No; 1 or 0; c 1 or c 2 Probability distribution: P(C) Probability value: P(C=c 1 ) or P(c 1 ) or P(c) P(A,B,C,W,D) P(a,b,c,w,d) full joint probability distribution some particular entry in the joint 15

A BN represents the full Joint (it contains all of the information in the Joint*) The elements of the joint are recoverable from the BN P( x n 1, x2, x3,... x ) n i 1 P( x i parents ( X i )) (eqn 14.2) The Joint is factorable Can we represent all Joints over X 1, X 2, X n? Which cannot be represented in the BN? * Provided the conditional independences hold 16

How flexible is The Joint vs. A Bayesian Net Q: What Joints cannot be represented? A: Those that violate the BNs conditional independence assumptions Q: How is a conditional independence assumption denoted in a BN? A: The lack of an arc. 17

A Graphical Model (for us in statistical AI) A (compact) representation of a particular statistical distribution Structural or graphical Part: Commitment to random variables Observable Latent (Unobservable) Captures restrictions on how they can interact Denotes a family of distributions Parametric Part: Set of parameters (sufficient statistics) Their values individuate a distribution from the family 18