Bayesian Networks. Machine Learning, Fall Slides based on material from the Russell and Norvig AI Book, Ch. 14

Similar documents
Directed Graphical Models

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

CS 343: Artificial Intelligence

PROBABILISTIC REASONING SYSTEMS

School of EECS Washington State University. Artificial Intelligence

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Probabilistic Reasoning Systems

Artificial Intelligence Bayes Nets: Independence

Bayesian networks. Chapter 14, Sections 1 4

Objectives. Probabilistic Reasoning Systems. Outline. Independence. Conditional independence. Conditional independence II.

Intelligent Systems (AI-2)

Bayesian Networks BY: MOHAMAD ALSABBAGH

Informatics 2D Reasoning and Agents Semester 2,

Bayes Nets: Independence

Bayesian networks. Chapter Chapter

Inference in Bayesian Networks

CS 5522: Artificial Intelligence II

Review: Bayesian learning and inference

Probabilistic Graphical Networks: Definitions and Basic Results

Bayesian Networks. Motivation

Bayesian Networks. Philipp Koehn. 6 April 2017

Graphical Models - Part I

Bayes Nets: Sampling

Bayesian Networks. Philipp Koehn. 29 October 2015

COS402- Artificial Intelligence Fall Lecture 10: Bayesian Networks & Exact Inference

Another look at Bayesian. inference

EE562 ARTIFICIAL INTELLIGENCE FOR ENGINEERS

Introduction to Artificial Intelligence Belief networks

Introduction to Artificial Intelligence. Unit # 11

Bayesian networks. Independence. Bayesian networks. Markov conditions Inference. by enumeration rejection sampling Gibbs sampler

Bayesian networks. Chapter AIMA2e Slides, Stuart Russell and Peter Norvig, Completed by Kazim Fouladi, Fall 2008 Chapter 14.

Learning Bayesian Networks (part 1) Goals for the lecture

Informatics 2D Reasoning and Agents Semester 2,

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech

CS 380: ARTIFICIAL INTELLIGENCE UNCERTAINTY. Santiago Ontañón

Artificial Intelligence Bayesian Networks

Foundations of Artificial Intelligence

Quantifying uncertainty & Bayesian networks

Bayesian Belief Network

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Bayesian Network. Outline. Bayesian Network. Syntax Semantics Exact inference by enumeration Exact inference by variable elimination

Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation

CS 188: Artificial Intelligence. Bayes Nets

CS 484 Data Mining. Classification 7. Some slides are from Professor Padhraic Smyth at UC Irvine

Bayes Networks 6.872/HST.950

Lecture 10: Bayesian Networks and Inference

Bayesian Networks Representation

Outline. CSE 573: Artificial Intelligence Autumn Bayes Nets: Big Picture. Bayes Net Semantics. Hidden Markov Models. Example Bayes Net: Car

Learning With Bayesian Networks. Markus Kalisch ETH Zürich

Uncertainty and Bayesian Networks

Stochastic inference in Bayesian networks, Markov chain Monte Carlo methods

14 PROBABILISTIC REASONING

Bayesian Networks. instructor: Matteo Pozzi. x 1. x 2. x 3 x 4. x 5. x 6. x 7. x 8. x 9. Lec : Urban Systems Modeling

Bayes Nets III: Inference

Bayesian networks: Modeling

Bayesian Networks (Part II)

Bayesian Networks. Material used

Reasoning Under Uncertainty: Belief Network Inference

Probabilistic Reasoning. Kee-Eung Kim KAIST Computer Science

Bayesian Networks. Semantics of Bayes Nets. Example (Binary valued Variables) CSC384: Intro to Artificial Intelligence Reasoning under Uncertainty-III

COMP9414: Artificial Intelligence Reasoning Under Uncertainty

CS Lecture 3. More Bayesian Networks

COMP5211 Lecture Note on Reasoning under Uncertainty

Bayes Networks. CS540 Bryan R Gibson University of Wisconsin-Madison. Slides adapted from those used by Prof. Jerry Zhu, CS540-1

Bayesian networks. Chapter Chapter

Course Overview. Summary. Outline

Bayesian networks. Chapter Chapter Outline. Syntax Semantics Parameterized distributions. Chapter

Sampling from Bayes Nets

Lecture 6: Graphical Models

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.

last two digits of your SID

Belief Networks for Probabilistic Inference

Directed Graphical Models or Bayesian Networks

Probability. CS 3793/5233 Artificial Intelligence Probability 1

Machine Learning for Data Science (CS4786) Lecture 24

Bayesian Networks Introduction to Machine Learning. Matt Gormley Lecture 24 April 9, 2018

CPSC 540: Machine Learning

Bayesian networks. Soleymani. CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018

Rapid Introduction to Machine Learning/ Deep Learning

This lecture. Reading. Conditional Independence Bayesian (Belief) Networks: Syntax and semantics. Chapter CS151, Spring 2004

Outline } Conditional independence } Bayesian networks: syntax and semantics } Exact inference } Approximate inference AIMA Slides cstuart Russell and

CS Belief networks. Chapter

ECE 6504: Advanced Topics in Machine Learning Probabilistic Graphical Models and Large-Scale Learning

Bayesian belief networks. Inference.

Approximate Inference

Uncertainty. Logic and Uncertainty. Russell & Norvig. Readings: Chapter 13. One problem with logical-agent approaches: C:145 Artificial

Probabilistic Models. Models describe how (a portion of) the world works

Lecture 8: Bayesian Networks

Uncertainty and Belief Networks. Introduction to Artificial Intelligence CS 151 Lecture 1 continued Ok, Lecture 2!

Uncertainty. 22c:145 Artificial Intelligence. Problem of Logic Agents. Foundations of Probability. Axioms of Probability

Introduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013

Events A and B are independent P(A) = P(A B) = P(A B) / P(B)

Conditional Independence

Axioms of Probability? Notation. Bayesian Networks. Bayesian Networks. Today we ll introduce Bayesian Networks.

Product rule. Chain rule

Probabilistic Representation and Reasoning

Intelligent Systems (AI-2)

Probabilistic Graphical Models

Reasoning Under Uncertainty

Transcription:

Bayesian Networks Machine Learning, Fall 2010 Slides based on material from the Russell and Norvig AI Book, Ch. 14 1

Administrativia Bayesian networks The inference problem: given a BN, how to make predictions based on it The learning problem: given a data set D, how to learn a BN that gives the best fit to D No class on Thursday Office hours: Today 12:30-1:30; Wed. cancelled 2

Bayesian Network Directed Acyclic Graph Overview Each node corresponds to one of the variables (attributes) that describe instances Edges indicate (causal) relationships Each node X i maintains the conditional probability distribution P(X i Parents(X i )) A joint assignment x 1,x 2,... x n Compact representation of a joint probability distribution over a set of variables. to the variables X 1,X 2,... X n has probability P ( x 1,x 2,... x n )= 3 n i=1 P (x i Parents(x i ))

Classic BN Example Picture from Russell and Norvig AI Book 4

Conditional Independence in a BN X is conditionally independent of its nondescendants given its parents X is conditionally independent of all other nodes given its parents, its children, and it s children s parents also known as its Markov blanket 5

Predicting with a BN Called inference Given: A BN Observations for some of the variables What can be inferred about the unobserved variables? 6

Probability Cheat Sheet Before talking about inference, let s review some probability basics Given a joint distribution over two Boolean variables, A and B, how can we use it to derive P(A B)? P (A, B) P (A B) = = P (A, B) P (B) To find what α is, simply compute P (A B) =αp (A, B) and P ( A B) =αp ( A, B) and normalize 7

Cheat Sheet Continued P (A B) = Given a joint distribution over three Boolean variables A, B, and C, how can we find P(A B)? P (A B, C = c) = c Values(C) c Values(C) P (A, B, C = c) What if we have A, B, C, and D and want P(A B)? Sum over joint assignments to values of C and D 8

Exact Inference in BNs Want to answer questions like What is the probability of one of the variables, given values for some (not necessarily all) of the others E.g. in Alarm network, what is P(Burglary JohnCalls = t, MaryCalls = t)? P (B = t J = t, M = t) =α According to our cheat sheet, this is e Values(E) a Values(A) P (B = t, J = t, M = t, E = e, A = a) 9

Variable Elimination Works by summing out (eliminating) uninteresting variables one at a time, doing so in a smart way to reuse computation.... not going to cover it (or any other exact algo) in detail... because in the worst case exact inference in BNs is an NP-hard problem special cases are tractable 10

Approximate Inference in BNs Based on sampling Will use this example from Russell and Norvig AI Book: 11

Direct Sampling First, assume the values of none of the variables are given as evidence e.g., want P(WetGrass) Simple sampling procedure: How to generate samples from a BN? Taking the generative approach, analogous to how we did it in Naive Bayes Generate N = large number of samples from the BN P(WetGrass) is computed as the number of samples in which WetGrass was t, out of all generated samples 12

Rejection Sampling What if we have observed some of the variables as evidence? i.e., want P(WetGrass Cloudy) Take same approach as before, but reject samples in which Cloudy is not set to true Compute counts based on just un-rejected samples Problem: We may end up rejecting most of our samples! 13

MCMC Sampling Rather than generating each sample from scratch, why not use previous sample to generate the next one? 14

MCMC Basic Idea Set observed variables to their observed values Set unobserved variables to random values Begin sampling: Generate a new value for an unobserved variable X, given the currently set values of the variables in its Markov blanket 15

Sampling Given the Markov Blanket We only know how to sample a value for X given it s parents values However, MB(X) includes X s parents, children, and the other parents of its children So, how to compute P(X MB(X))? P (x MB(X)) = αp (x Parents(X)) P (y Parents(Y )) Y Chldrn(X) 16