Artificial Intelligence Bayes Nets: Independence

Similar documents
Bayes Nets: Independence

CS 5522: Artificial Intelligence II

Product rule. Chain rule

Recap: Bayes Nets. CS 473: Artificial Intelligence Bayes Nets: Independence. Conditional Independence. Bayes Nets. Independence in a BN

Probabilistic Models. Models describe how (a portion of) the world works

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation

CSE 473: Artificial Intelligence Autumn 2011

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

Bayes Nets. CS 188: Artificial Intelligence Fall Example: Alarm Network. Bayes Net Semantics. Building the (Entire) Joint. Size of a Bayes Net

Bayesian Networks. Vibhav Gogate The University of Texas at Dallas

Announcements. CS 188: Artificial Intelligence Fall Example Bayes Net. Bayes Nets. Example: Traffic. Bayes Net Semantics

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018

Outline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

Outline. CSE 473: Artificial Intelligence Spring Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

CS 188: Artificial Intelligence Spring Announcements

CS 188: Artificial Intelligence Spring Announcements

CS 343: Artificial Intelligence

CS 5522: Artificial Intelligence II

Bayesian networks (1) Lirong Xia

CS 343: Artificial Intelligence

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

CS 188: Artificial Intelligence Fall 2008

CSEP 573: Artificial Intelligence

CS 188: Artificial Intelligence Fall 2009

Announcements. CS 188: Artificial Intelligence Spring Bayes Net Semantics. Probabilities in BNs. All Conditional Independences

Bayes Nets III: Inference

Probabilistic Models

CS 188: Artificial Intelligence. Bayes Nets

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian networks. Chapter 14, Sections 1 4

CSE 473: Ar+ficial Intelligence. Hidden Markov Models. Bayes Nets. Two random variable at each +me step Hidden state, X i Observa+on, E i

Directed Graphical Models

CS 188: Artificial Intelligence

Bayesian networks. Chapter Chapter

CSE 473: Artificial Intelligence Probability Review à Markov Models. Outline

Introduction to Artificial Intelligence Belief networks

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

Bayesian Network. Outline. Bayesian Network. Syntax Semantics Exact inference by enumeration Exact inference by variable elimination

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech

Bayes Net Representation. CS 188: Artificial Intelligence. Approximate Inference: Sampling. Variable Elimination. Sampling.

Introduction to Artificial Intelligence. Unit # 11

PROBABILISTIC REASONING SYSTEMS

COS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

Objectives. Probabilistic Reasoning Systems. Outline. Independence. Conditional independence. Conditional independence II.

Bayesian Networks. Machine Learning, Fall Slides based on material from the Russell and Norvig AI Book, Ch. 14

Probabilistic Reasoning Systems

last two digits of your SID

Uncertainty and Bayesian Networks

Bayes Nets: Sampling

Quantifying uncertainty & Bayesian networks

Bayesian Networks. Semantics of Bayes Nets. Example (Binary valued Variables) CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-III

CS 343: Artificial Intelligence

CSE 473: Artificial Intelligence

Probability. CS 3793/5233 Artificial Intelligence Probability 1

CS188 Outline. We re done with Part I: Search and Planning! Part II: Probabilistic Reasoning. Part III: Machine Learning

CS 343: Artificial Intelligence

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.

CS 380: ARTIFICIAL INTELLIGENCE UNCERTAINTY. Santiago Ontañón

Intelligent Systems (AI-2)

Directed Graphical Models or Bayesian Networks

Bayesian networks. Chapter AIMA2e Slides, Stuart Russell and Peter Norvig, Completed by Kazim Fouladi, Fall 2008 Chapter 14.

CS 188: Artificial Intelligence Spring Announcements

Machine Learning Lecture 14

Bayes Networks 6.872/HST.950

Review: Bayesian learning and inference

CS 343: Artificial Intelligence

Informatics 2D Reasoning and Agents Semester 2,

Graphical Models - Part I

Introduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013

Bayes Networks. CS540 Bryan R Gibson University of Wisconsin-Madison. Slides adapted from those used by Prof. Jerry Zhu, CS540-1

Bayesian networks: Modeling

CMPSCI 240: Reasoning about Uncertainty

Bayesian networks. Chapter Chapter

Probabilistic Reasoning. Kee-Eung Kim KAIST Computer Science

Y. Xiang, Inference with Uncertain Knowledge 1

qfundamental Assumption qbeyond the Bayes net Chain rule conditional independence assumptions,

CS 484 Data Mining. Classification 7. Some slides are from Professor Padhraic Smyth at UC Irvine

CS188 Outline. CS 188: Artificial Intelligence. Today. Inference in Ghostbusters. Probability. We re done with Part I: Search and Planning!

CS 188: Artificial Intelligence. Our Status in CS188

Bayesian networks. Chapter Chapter Outline. Syntax Semantics Parameterized distributions. Chapter

CS 5522: Artificial Intelligence II

CS Lecture 3. More Bayesian Networks

CS 343: Artificial Intelligence

Bayesian Networks. Motivation

Directed Graphical Models

Approximate Inference

Artificial Intelligence Bayesian Networks

School of EECS Washington State University. Artificial Intelligence

Our Status. We re done with Part I Search and Planning!

15-780: Graduate Artificial Intelligence. Bayesian networks: Construction and inference

CS 730/730W/830: Intro AI

Extensions of Bayesian Networks. Outline. Bayesian Network. Reasoning under Uncertainty. Features of Bayesian Networks.

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning

Graphical Models and Kernel Methods

Introduction to Probabilistic Reasoning. Image credit: NASA. Assignment

CS Belief networks. Chapter

Reinforcement Learning Wrap-up

Transcription:

Artificial Intelligence Bayes Nets: Independence Instructors: David Suter and Qince Li Course Delivered @ Harbin Institute of Technology [Many slides adapted from those created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. Some others from colleagues at Adelaide University.]

Probability Recap Conditional probability Product rule Chain rule X, Y independent if and only if: X and Y are conditionally independent given Z if and only if: Product of parents cond on child

Bayes Nets A Bayes net is an efficient encoding of a probabilistic model of a domain Questions we can ask: Inference: given a fixed BN, what is P(X e)? Representation: given a BN graph, what kinds of distributions can it encode? Modeling: what BN is most appropriate for a given domain?

Bayes Net Semantics A directed, acyclic graph, one node per random variable A conditional probability table (CPT) for each node A collection of distributions over X, one for each combination of parents values Bayes nets implicitly encode joint distributions As a product of local conditional distributions To see what probability a BN gives to a full assignment, multiply all the relevant conditionals together:

Example: Alarm Network B P(B) +b 0.001 B E E P(E) +e 0.002 -b 0.999 -e 0.998 A J P(J A) +a +j 0.9 +a -j 0.1 -a +j 0.05 -a -j 0.95 J A M A M P(M A) +a +m 0.7 +a -m 0.3 -a +m 0.01 -a -m 0.99 B E A P(A B,E) +b +e +a 0.95 +b +e -a 0.05 +b -e +a 0.94 +b -e -a 0.06 -b +e +a 0.29 -b +e -a 0.71 -b -e +a 0.001 -b -e -a 0.999

Example: Alarm Network B P(B) +b 0.001 B E E P(E) +e 0.002 -b 0.999 -e 0.998 A J P(J A) +a +j 0.9 +a -j 0.1 -a +j 0.05 -a -j 0.95 J A M A M P(M A) +a +m 0.7 +a -m 0.3 -a +m 0.01 -a -m 0.99 B E A P(A B,E) +b +e +a 0.95 +b +e -a 0.05 +b -e +a 0.94 +b -e -a 0.06 -b +e +a 0.29 -b +e -a 0.71 -b -e +a 0.001 -b -e -a 0.999

Size of a Bayes Net How big is a joint distribution over N Boolean variables? 2 N How big is an N-node net if nodes have up to k parents? O(N * 2 k+1 ) Both give you the power to calculate BNs: Huge space savings! Also easier to elicit local CPTs Also faster to answer queries (coming)

Bayes Nets Representation Conditional Independences Probabilistic Inference Learning Bayes Nets from Data

Conditional Independence X and Y are independent if X and Y are conditionally independent given Z (Conditional) independence is a property of a distribution Example:

Independence in a BN Important question about a BN: Are two nodes independent given certain evidence? If yes, can prove using algebra (tedious in general) If no, can prove with a counter example Example: X Y Z Question: are X and Z necessarily independent? Answer: no. Example: low pressure causes rain, which causes traffic. X can influence Z, Z can influence X (via Y) Addendum: they could be independent: how?

D-separation: Outline

D-separation: Outline Study independence properties for triples Analyze complex cases in terms of member triples D-separation: a condition / algorithm for answering such queries

Causal Chains This configuration is a causal chain Guaranteed X independent of Z? No! One example set of CPTs for which X is not independent of Z is sufficient to show this independence is not guaranteed. Example: X: Low pressure Y: Rain Z: Traffic Low pressure causes rain causes traffic, high pressure causes no rain causes no traffic In numbers: P( +y +x ) = 1, P( -y - x ) = 1, P( +z +y ) = 1, P( -z -y ) = 1

Causal Chains This configuration is a causal chain Guaranteed X independent of Z given Y? X: Low pressure Y: Rain Z: Traffic Yes! Evidence along the chain blocks the influence

Common Cause This configuration is a common cause Y: Project due Guaranteed X independent of Z? No! One example set of CPTs for which X is not independent of Z is sufficient to show this independence is not guaranteed. X: Forums busy Z: Lab full Example: Project due causes both forums busy and lab full In numbers: P( +x +y ) = 1, P( -x -y ) = 1, P( +z +y ) = 1, P( -z -y ) = 1

Common Cause This configuration is a common cause Y: Project due Guaranteed X and Z independent given Y? X: Forums busy Z: Lab full Yes! Observing the cause blocks influence between effects.

Common Effect Last configuration: two causes of one effect (v-structures) X: Raining Y: Ballgame Are X and Y independent? Yes: the ballgame and the rain cause traffic, but they are not correlated Still need to prove they must be (try it!) Are X and Y independent given Z? No: seeing traffic puts the rain and the ballgame in competition as explanation. Z: Traffic This is backwards from the other cases Observing an effect activates influence between possible causes.

Independence in Larger Than Triplet The fundamental design decision in representing a model as a Bayes Net/Graphical Model is that the variable of a node is conditionally independent of its non-descendents given its parents. This is exactly the step from the general factorization: To the special factorization: 18 To see this work your way up from the bottom of the network...

Independence larger than triples A variable of a node is conditionally independent of all other variables (of other nodes), given its parents, its children and its childrens parents. That is, given its Markov Blanket. (see AIMA 3 rd ed. Section 2.2). 19

Independence larger than triples 20

The General Case D-separation is more general.the following may be of use in some problems...

The General Case General question: in a given BN, are two variables independent (given evidence)? Solution: analyze the graph Any complex example can be broken into repetitions of the three canonical cases

Reachability Recipe: shade evidence nodes, look for paths in the resulting graph Attempt 1: if two nodes are connected by an undirected path not blocked by a shaded node, they are conditionally independent D L R T B Almost works, but not quite Where does it break? Answer: the v-structure at T doesn t count as a link in a path unless active

Active / Inactive Paths Question: Are X and Y conditionally independent given evidence variables {Z}? Yes, if X and Y d-separated by Z Consider all (undirected) paths from X to Y No active paths = independence! Active Triples Inactive Triples A path is active if each triple is active: Causal chain A B C where B is unobserved (either direction) Common cause A B C where B is unobserved Common effect (aka v-structure) A B C where B or one of its descendents is observed All it takes to block a path is a single inactive segment

D-Separation Query: X i X j {X k1,...,x kn }? Check all (undirected!) paths between and If one or more active, then independence not guaranteed X i X j {X k1,...,x kn } Otherwise (i.e. if all paths are inactive), then independence is guaranteed X i X j {X k1,...,x kn }

Example Yes R B T T

Example Yes Yes L R B D T Yes T

Example Variables: R: Raining T: Traffic D: Roof drips S: I m sad Questions: T R S D Yes

Bayes Nets Representation Summary Bayes nets compactly encode joint distributions Guaranteed independencies of distributions can be deduced from BN graph structure D-separation gives precise conditional independence guarantees from graph alone A Bayes net s joint distribution may have further (conditional) independence that is not detectable until you inspect its specific distribution

Bayes Nets Representation Conditional Independences Probabilistic Inference Enumeration (exact, exponential complexity) Variable elimination (exact, worst-case exponential complexity, often better) Probabilistic inference is NP-complete Sampling (approximate) Learning Bayes Nets from Data

31