Bayesian Networks Representation and Reasoning

Size: px
Start display at page:

Download "Bayesian Networks Representation and Reasoning"

Transcription

1 ayesian Networks Representation and Reasoning Marco F. Ramoni hildren s Hospital Informatics Program Harvard Medical School (2003) Harvard-MIT ivision of Health Sciences and Technology HST.951J: Medical ecision Support

2 Introduction ayesian network are a knowledge representation formalism for reasoning under uncertainty. ayesian network is a direct acyclic graph encoding assumptions of conditional independence. In a ayesian network, nodes are stochastic variables and arcs are dependency between nodes. ayesian networks were designed to encode explicitly encode deep knowledge rather than heuristics, to simplify knowledge acquisition, provide a firmer theoretical ground, and foster reusability.

3 Graph graph (network) G(N,L) is defined by: Nodes: finite set N = {,,...} of nodes (vertices). rcs: set L of arcs (edges): ordered pair of nodes. Set L is a subset of all possible pairs of nodes N. L={(,),(,),(,)} L={(,),(,)} L={(,),(,),(,),(,),(,),(,)}

4 Graph irected Types of Graph Undirected onnected isconnected cyclic yclic Singly onnected (Tree) Simple Tree Polytree Multiply onnected

5 irection irection of a link: irected: if (,) is in N, then (,) is not in N. Undirected: if (,) is in N, then (,) is in N. Note: The link should be «. haracters: djacent set: the nodes one step away from : dj()={ (,) L}. Path: The set of n nodes X i from to via links: Loop: closed path: X 1 = X n. cyclic graph: graph with no cycles.

6 irected Graphs Parent: is parent of if there is a directed link fi. Family: The set made up by a node and its parents. ncestor: is ancestor of if exists a path from to. ncestral set: set of nodes containing their ancestors. ycle: cycle is a closed loop of directed links. ssociated acyclic graph: The undirected graph obtained by dropping the direction of links. Moral graph: The undirected graph obtained by. Marring the parents of a common child. ropping the directions of the links.

7 Trees Tree: If every pair of nodes there is at most one path. Simple Tree: ach node has at most one parent. PolyTree: Nodes can have more than one parent. Multiply onnected Graph: graph where at least one pair of nodes has more than one path. Note: The associated undirected graph has a loop.

8 ayesian Networks Qualitative: dependency graph made by: Node: a variable X, with a set of states {x 1,,x n }. rc: a dependency of a variable X on its parents P. Quantitative: The distributions of a variable X given each combination of states p i of its parents P. p() p() Y Y O O p() p() L L H H I =ge ge; ; =ducation ducation; ; I=Income Income I I p(i,) p(i,) Y Y L L L L Y Y L L H H Y Y H H L L Y Y H H H H O O L L L L O O L L H H O O H H L L O O H H H H

9 Independence Perfect dependence between isease and Test: Test isease Test isease Independence between isease and Test: Test isease Test isease xercise: ompute the PT for Test given isease.

10 Why o We are? Independence simplifies models: if two variables are independent, I do not need to model their interaction but I can reason about them separately. In this form of independence, called marginal independence, however, a variable will tell me nothing about another variable, by design. There is another, more useful, form of independence, which maintains the connection between variables but, at the same time, breaks down the whole system in separate regions: conditional independence. This is independence used by ayesian networks.

11 onditional Independence When two variables are independent given a third, they are said to be conditionally independent. p( )=p( )/p( )=p( ). Literacy Literacy T-shirt Yes No Small Large ge T-shirt Yes No <5 Small <5 Large >5 Small >5 Large

12 ayesian Networks ayesian networks use graphs to capture these statement of conditional independence. ayesian network (N) is defined by a graph: Nodes are stochastic variables. Links are dependencies. No link means independence given a parent. There are two components in a N: Qualitative graphical structure. Quantitative assessment of probabilities. ge T-shirt Literacy

13 ecomposition Ns decompose the joint probability distribution with the graph of conditional independence. Therefore, the graphical structure factorizes the joint probability distribution: p( ) = p( ) p( ) p().

14 Markov quivalence ifferent network structures may encode the same conditional independence statements: and are conditionally independent given. can be encoded by 3 different network structures. In all these network structures, the information flow running between and along the direction of the arrows is mediated by the node.

15 xample ackground knowledge: General rules of behavior. p(ge=<5)=0.3 p(t-shirt=small ge=<5)=0.5 p(t-shirt=small ge=>5)=0.3 p(literacy=yes ge=>5)=0.6 p(literacy=yes ge=<5)=0.2. vidence: Observation p(t-shirt=small). Solution: The posterior probability distribution of the unobserved nodes given evidence: p(literacy T-shirt=small) and p(ge T-shirt=small). p(ge=<5,t-shirt=small,literacy=yes) p(ge=<5,t-shirt=small,literacy=no) p(ge=<5,t-shirt=large,literacy=yes) ge p(ge=<5,t-shirt=large,literacy =no) p(ge=>5,t-shirt=small,literacy=yes) p(ge=>5,t-shirt=small,literacy=no) p(ge=>5,t-shirt=large,literacy=yes) p(ge=>5,t-shirt=large, Literacy=no). T-shirt Literacy

16 Reasoning omponents of a problem: Knowledge: graph and numbers. vidence: e={c and g}. Solution: p(d c,g)=? Note: Lower case is an instance. F G p() p() p( ) F p() p(f ) p(,) G p(g,)

17 ompute the Joint Probability istribution: rute Force p(a,b,c,d,e,f,g)=p(a)p(b)p(c d)p(d a,b)p(e)p(f d)p(g d,e). Marginalize out the variable of interest: p(d)=s p(a,b,c,e,f,g). Note: We have replaced with, ost: 2 n probabilities (2 6 = 64)

18 ecomposition ecomposition: breaks the N into two Ns: p(d)= S p(a)p(b)p(c a)p(d a,b)p(e)p(f d)p(g d,e)=. = (S p(a)p(b)p(c a)p(d a,b)) (S p(e)p(f d)p(g d,e)). Saving: We move from 64 to =16, and most of all the terms move from 7 to 4 and from 7 to 3. -separation: the basic idea is based on a property of graphs called d-separation (directed-separation).

19 Propagation in Polytrees In a polytree, each node breaks the graph into two independent sub-graphs and evidence can be independently propagated in the two graphs: +: evidence coming from the parents (+ = {c}). -: evidence coming from the children (- = {g}). F G

20 Message Passing Message passing algorithm (Kim & Pearl 1983) is a local propagation method for polytrees. The basic idea is that p(d) is actually made up by parent component p(d) and a south component l(d). The basic idea is to loop and pass p and l messages between nodes until no message can be passed. In this way, the propagation is entirely distributed and the computations are locally executed in each node.

21 lgorithm Input: N with a set of variables X and a set of evidential statements e = {=a,=b, }. Output: onditional probability distribution p(x e) for each non evidential variable X. Initialization Step: ach evidential variable X, if x e p(x)=1, else p(x)=0. if x e l(x)=1, else l(x)=0. ach non evidential root variable X, p(x) = p(x). ach non evidential childless variable X, l(x)=1.

22 lgorithm II Iteration Step (on non evidential variables X/e): If X has all the p-messages from its parents, p(x). If X has all the l-messages from its children, l(x). If p(x), for each child, if l -messages from all other children are in, send p-message to child. If l(x), for each parent, if p-messages from all other parents are in, send l-message to parent. Repeat until no message is sent. losure: For each X/e, compute b(x)= p(x) l(x). For each X/e, compute p(x)= b(x)/s l(x i ).

23 Properties istributed: ach node does not need to know about the others when it is passing the information around. Parallel architecture: ach node can be imagined as a separate processor. omplexity: Linear in the number of nodes. Limitations: onfined to a restricted class of graphs and, most of all, unable to represent an important class of problems. Importance: Proof of feasibility - ayesians are not just dreamers after all.

24 Multiply onnected N When the N is a multiply connected graph. The associated undirected graph contains a loop. ach node does not break the network into 2 parts. Information may flow through more than one paths. Pearl s lgorithm is no longer applicable. F G F G

25 Main stream methods: onditioning Methods. lustering Methods. Methods The basic strategy is: Turn multiply connected graph in something else. Use Pearl s algorithm to propagate evidence. Recover the conditional probability p(x e) for X. Methods differ in the way in which. What they transform the graph into. The properties they exploit for this transformation.

26 onditioning Methods The transformation strategy is: Instantiate a set of nodes (cutset) to cut the loops. bsorb evidence and change the graph topology. Propagate each N using Pearl s algorithm. Marginalize with respect to the loop cutset. =e F G F G

27 lgorithm Input: a (multiply connected) N and evidence e. Output: the posterior probability p(x e) for each X. Procedure: 1. Identify a loop cutset =( 1,, n ). 2. For each member of combinations c=(c 1,, c n ). Generate a polytree Ns for each c. Use Pearl s lgorithm to compute p(x e,c 1,,c n ). ompute p(c 1,, c n e) = p(e c 1,,c n )p(c 1,,c n ) /p(e). 3. For each node X, a=p(x e) S c p(x e,c 1,,c n )p(e c 1,,c n )p(c 1,,c n ), ompute p(x e)= a/s x p(x).

28 omplexity The computational complexity is exponential in the size of the loop cutset, as we must generate and propagate a N for each combination of states of the loop cutset. The identification of the minimal loop cutset of a N is NP-hard, but heuristic methods exist to make it feasible. The computational complexity is a problem common to all methods moving from polytrees to multiply connected graphs.

29 xample Multiply connected N No evidence p() p( ) p( ) p( ) F p(f ) F p(,)

30 Loop cutset: {}. p(=0)=p(=0 =0)p(=1) + p(=0 =1)p(=1) xample F F F

31 lustering Methods The basic strategy (Lauritzen & Spiegelhalter 1988) is: 1. onvert a N in a undirected graph coding the same conditional independence assumptions. 2. nsure the resulting graph is decomposable. 3. This operation clusters nodes in locally independent subgraphs (cliques). 4. These cliques are joint to each other via a single nodes. 5. Produce a perfect numbering of nodes. 6. Recursively propagate evidence.

32 Markov Networks Markov network is a based on undirected graphs: N : G = Markov Network : Undirected Graph. Markov networks encode conditional independence assumptions (as Ns) using a Undirected Graph: 1. link between and means dependency. 2. variable is independent of all not adjacent variables given the adjacent ones. xample: is independent from (,,) given.

33 ecomposable ecomposable Markov networks lead to efficiency: Markov network is said to be decomposable when it contains no cycle with longer than 3 (there is no unbroken cycle with more than 3 nodes). The joint probability distribution of the graph can be factorized by the marginal distributions of the cliques: clique is the largest sub-graph in which nodes are all adjacent to each other. Therefore, a clique cannot be further simplified by conditional independence assumptions.

34 Triangulation When a Markov network is not decomposable, we triangulate the graph by including the missing links. The product of the joint probability of each clique, divided by the product of their intersection: p(a,b,c)=p(c a)p(b a)p(a). Moralize 1.Marry parents 2.rop arrows

35 Reading Independence The translation method via moralization reads the conditional independence statements in N. Gs cannot encode any arbitrary set of conditional independence assumptions. I(, (,)) I(, (,))

36 Propagation ompile the N into a moralized Markov network. Maximum cardinality search: For each clique Q compute p(q e). Within each cluster, marginalize p(x e)

37 Who is the Winner? lustering is also NP-complete. The source of computational complexity is the size of the larger clique in the graph. Global conditioning (Shachter, ndersen & Szolovits 1994) shows that: 1. onditioning is a special case of lustering. 2. onditioning is better at trading off memory-time. 3. onditioning is better suited for parallel implementations.

COMPSCI 276 Fall 2007

COMPSCI 276 Fall 2007 Exact Inference lgorithms for Probabilistic Reasoning; OMPSI 276 Fall 2007 1 elief Updating Smoking lung ancer ronchitis X-ray Dyspnoea P lung cancer=yes smoking=no, dyspnoea=yes =? 2 Probabilistic Inference

More information

Machine Learning Lecture 14

Machine Learning Lecture 14 Many slides adapted from B. Schiele, S. Roth, Z. Gharahmani Machine Learning Lecture 14 Undirected Graphical Models & Inference 23.06.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de

More information

Discrete Bayesian Networks: The Exact Posterior Marginal Distributions

Discrete Bayesian Networks: The Exact Posterior Marginal Distributions arxiv:1411.6300v1 [cs.ai] 23 Nov 2014 Discrete Bayesian Networks: The Exact Posterior Marginal Distributions Do Le (Paul) Minh Department of ISDS, California State University, Fullerton CA 92831, USA dminh@fullerton.edu

More information

Undirected Graphical Models 4 Bayesian Networks and Markov Networks. Bayesian Networks to Markov Networks

Undirected Graphical Models 4 Bayesian Networks and Markov Networks. Bayesian Networks to Markov Networks Undirected Graphical Models 4 ayesian Networks and Markov Networks 1 ayesian Networks to Markov Networks 2 1 Ns to MNs X Y Z Ns can represent independence constraints that MN cannot MNs can represent independence

More information

Message Passing Algorithms and Junction Tree Algorithms

Message Passing Algorithms and Junction Tree Algorithms Message Passing lgorithms and Junction Tree lgorithms Le Song Machine Learning II: dvanced Topics S 8803ML, Spring 2012 Inference in raphical Models eneral form of the inference problem P X 1,, X n Ψ(

More information

Inference in Graphical Models Variable Elimination and Message Passing Algorithm

Inference in Graphical Models Variable Elimination and Message Passing Algorithm Inference in Graphical Models Variable Elimination and Message Passing lgorithm Le Song Machine Learning II: dvanced Topics SE 8803ML, Spring 2012 onditional Independence ssumptions Local Markov ssumption

More information

Machine Learning Summer School

Machine Learning Summer School Machine Learning Summer School Lecture 1: Introduction to Graphical Models Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ epartment of ngineering University of ambridge, UK

More information

Probabilistic Graphical Models (I)

Probabilistic Graphical Models (I) Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random

More information

6.867 Machine learning, lecture 23 (Jaakkola)

6.867 Machine learning, lecture 23 (Jaakkola) Lecture topics: Markov Random Fields Probabilistic inference Markov Random Fields We will briefly go over undirected graphical models or Markov Random Fields (MRFs) as they will be needed in the context

More information

Exact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = =

Exact Inference I. Mark Peot. In this lecture we will look at issues associated with exact inference. = = Exact Inference I Mark Peot In this lecture we will look at issues associated with exact inference 10 Queries The objective of probabilistic inference is to compute a joint distribution of a set of query

More information

Statistical Approaches to Learning and Discovery

Statistical Approaches to Learning and Discovery Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University

More information

Bayesian Networks: Representation, Variable Elimination

Bayesian Networks: Representation, Variable Elimination Bayesian Networks: Representation, Variable Elimination CS 6375: Machine Learning Class Notes Instructor: Vibhav Gogate The University of Texas at Dallas We can view a Bayesian network as a compact representation

More information

Inference in Bayesian Networks

Inference in Bayesian Networks Andrea Passerini passerini@disi.unitn.it Machine Learning Inference in graphical models Description Assume we have evidence e on the state of a subset of variables E in the model (i.e. Bayesian Network)

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 16 Undirected Graphs Undirected Separation Inferring Marginals & Conditionals Moralization Junction Trees Triangulation Undirected Graphs Separation

More information

Chris Bishop s PRML Ch. 8: Graphical Models

Chris Bishop s PRML Ch. 8: Graphical Models Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular

More information

Lecture 17: May 29, 2002

Lecture 17: May 29, 2002 EE596 Pat. Recog. II: Introduction to Graphical Models University of Washington Spring 2000 Dept. of Electrical Engineering Lecture 17: May 29, 2002 Lecturer: Jeff ilmes Scribe: Kurt Partridge, Salvador

More information

Undirected Graphical Models

Undirected Graphical Models Undirected Graphical Models 1 Conditional Independence Graphs Let G = (V, E) be an undirected graph with vertex set V and edge set E, and let A, B, and C be subsets of vertices. We say that C separates

More information

Bayesian Networks Learning From Data

Bayesian Networks Learning From Data Bayesian Networks earning From Data Marco F. Ramoni hildren s ospital Informatics Program arvard Medical School ST951 (2003) arvard-mit Division of ealth Sciences and Technology ST.951J: Medical Decision

More information

Probability Propagation

Probability Propagation Graphical Models, Lecture 12, Michaelmas Term 2010 November 19, 2010 Characterizing chordal graphs The following are equivalent for any undirected graph G. (i) G is chordal; (ii) G is decomposable; (iii)

More information

CS281A/Stat241A Lecture 19

CS281A/Stat241A Lecture 19 CS281A/Stat241A Lecture 19 p. 1/4 CS281A/Stat241A Lecture 19 Junction Tree Algorithm Peter Bartlett CS281A/Stat241A Lecture 19 p. 2/4 Announcements My office hours: Tuesday Nov 3 (today), 1-2pm, in 723

More information

Implementing Machine Reasoning using Bayesian Network in Big Data Analytics

Implementing Machine Reasoning using Bayesian Network in Big Data Analytics Implementing Machine Reasoning using Bayesian Network in Big Data Analytics Steve Cheng, Ph.D. Guest Speaker for EECS 6893 Big Data Analytics Columbia University October 26, 2017 Outline Introduction Probability

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression

More information

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2016/2017 Lesson 13 24 march 2017 Reasoning with Bayesian Networks Naïve Bayesian Systems...2 Example

More information

Variable Elimination: Algorithm

Variable Elimination: Algorithm Variable Elimination: Algorithm Sargur srihari@cedar.buffalo.edu 1 Topics 1. Types of Inference Algorithms 2. Variable Elimination: the Basic ideas 3. Variable Elimination Sum-Product VE Algorithm Sum-Product

More information

15-780: Graduate Artificial Intelligence. Bayesian networks: Construction and inference

15-780: Graduate Artificial Intelligence. Bayesian networks: Construction and inference 15-780: Graduate Artificial Intelligence ayesian networks: Construction and inference ayesian networks: Notations ayesian networks are directed acyclic graphs. Conditional probability tables (CPTs) P(Lo)

More information

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network. ecall from last time Lecture 3: onditional independence and graph structure onditional independencies implied by a belief network Independence maps (I-maps) Factorization theorem The Bayes ball algorithm

More information

Probability Propagation

Probability Propagation Graphical Models, Lectures 9 and 10, Michaelmas Term 2009 November 13, 2009 Characterizing chordal graphs The following are equivalent for any undirected graph G. (i) G is chordal; (ii) G is decomposable;

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft

More information

Bayesian Networks. Example of Inference. Another Bayes Net Example (from Andrew Moore) Example Bayes Net. Want to know P(A)? Proceed as follows:

Bayesian Networks. Example of Inference. Another Bayes Net Example (from Andrew Moore) Example Bayes Net. Want to know P(A)? Proceed as follows: ayesian Networks ayesian network (N) is a graphical representation of the direct dependencies over a set of variables, together with a set of conditional probability tables quantifying the strength of

More information

Message Passing and Junction Tree Algorithms. Kayhan Batmanghelich

Message Passing and Junction Tree Algorithms. Kayhan Batmanghelich Message Passing and Junction Tree Algorithms Kayhan Batmanghelich 1 Review 2 Review 3 Great Ideas in ML: Message Passing Each soldier receives reports from all branches of tree 3 here 7 here 1 of me 11

More information

Chapter 8 Cluster Graph & Belief Propagation. Probabilistic Graphical Models 2016 Fall

Chapter 8 Cluster Graph & Belief Propagation. Probabilistic Graphical Models 2016 Fall Chapter 8 Cluster Graph & elief ropagation robabilistic Graphical Models 2016 Fall Outlines Variable Elimination 消元法 imple case: linear chain ayesian networks VE in complex graphs Inferences in HMMs and

More information

The Necessity of Bounded Treewidth for Efficient Inference in Bayesian Networks

The Necessity of Bounded Treewidth for Efficient Inference in Bayesian Networks The Necessity of Bounded Treewidth for Efficient Inference in Bayesian Networks Johan H.P. Kwisthout and Hans L. Bodlaender and L.C. van der Gaag 1 Abstract. Algorithms for probabilistic inference in Bayesian

More information

Directed and Undirected Graphical Models

Directed and Undirected Graphical Models Directed and Undirected Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Last Lecture Refresher Lecture Plan Directed

More information

Directed Graphical Models

Directed Graphical Models CS 2750: Machine Learning Directed Graphical Models Prof. Adriana Kovashka University of Pittsburgh March 28, 2017 Graphical Models If no assumption of independence is made, must estimate an exponential

More information

Variable Elimination: Algorithm

Variable Elimination: Algorithm Variable Elimination: Algorithm Sargur srihari@cedar.buffalo.edu 1 Topics 1. Types of Inference Algorithms 2. Variable Elimination: the Basic ideas 3. Variable Elimination Sum-Product VE Algorithm Sum-Product

More information

9 Forward-backward algorithm, sum-product on factor graphs

9 Forward-backward algorithm, sum-product on factor graphs Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous

More information

Axioms of Probability? Notation. Bayesian Networks. Bayesian Networks. Today we ll introduce Bayesian Networks.

Axioms of Probability? Notation. Bayesian Networks. Bayesian Networks. Today we ll introduce Bayesian Networks. Bayesian Networks Today we ll introduce Bayesian Networks. This material is covered in chapters 13 and 14. Chapter 13 gives basic background on probability and Chapter 14 talks about Bayesian Networks.

More information

Lecture 12: May 09, Decomposable Graphs (continues from last time)

Lecture 12: May 09, Decomposable Graphs (continues from last time) 596 Pat. Recog. II: Introduction to Graphical Models University of Washington Spring 00 Dept. of lectrical ngineering Lecture : May 09, 00 Lecturer: Jeff Bilmes Scribe: Hansang ho, Izhak Shafran(000).

More information

Methods of Exact Inference. Cutset Conditioning. Cutset Conditioning Example 1. Cutset Conditioning Example 2. Cutset Conditioning Example 3

Methods of Exact Inference. Cutset Conditioning. Cutset Conditioning Example 1. Cutset Conditioning Example 2. Cutset Conditioning Example 3 Methods of xact Inference Lecture 11: xact Inference means modelling all the dependencies in the data. xact Inference his is only computationally feasible for small or sparse networks. Methods: utset onditioning

More information

Inference in Bayes Nets

Inference in Bayes Nets Inference in Bayes Nets Bruce D Ambrosio Oregon State University and Robert Fung Institute for Decision Systems Research July 23, 1994 Summer Institute on Probablility in AI 1994 Inference 1 1 Introduction

More information

Bayes Networks 6.872/HST.950

Bayes Networks 6.872/HST.950 Bayes Networks 6.872/HST.950 What Probabilistic Models Should We Use? Full joint distribution Completely expressive Hugely data-hungry Exponential computational complexity Naive Bayes (full conditional

More information

Cutset sampling for Bayesian networks

Cutset sampling for Bayesian networks Cutset sampling for Bayesian networks Cutset sampling for Bayesian networks Bozhena Bidyuk Rina Dechter School of Information and Computer Science University Of California Irvine Irvine, CA 92697-3425

More information

Belief Update in CLG Bayesian Networks With Lazy Propagation

Belief Update in CLG Bayesian Networks With Lazy Propagation Belief Update in CLG Bayesian Networks With Lazy Propagation Anders L Madsen HUGIN Expert A/S Gasværksvej 5 9000 Aalborg, Denmark Anders.L.Madsen@hugin.com Abstract In recent years Bayesian networks (BNs)

More information

Bayesian Machine Learning - Lecture 7

Bayesian Machine Learning - Lecture 7 Bayesian Machine Learning - Lecture 7 Guido Sanguinetti Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh gsanguin@inf.ed.ac.uk March 4, 2015 Today s lecture 1

More information

p L yi z n m x N n xi

p L yi z n m x N n xi y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen

More information

Causal Discovery Methods Using Causal Probabilistic Networks

Causal Discovery Methods Using Causal Probabilistic Networks ausal iscovery Methods Using ausal Probabilistic Networks MINFO 2004, T02: Machine Learning Methods for ecision Support and iscovery onstantin F. liferis & Ioannis Tsamardinos iscovery Systems Laboratory

More information

A Tutorial on Bayesian Belief Networks

A Tutorial on Bayesian Belief Networks A Tutorial on Bayesian Belief Networks Mark L Krieg Surveillance Systems Division Electronics and Surveillance Research Laboratory DSTO TN 0403 ABSTRACT This tutorial provides an overview of Bayesian belief

More information

Part I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS

Part I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Part I C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Probabilistic Graphical Models Graphical representation of a probabilistic model Each variable corresponds to a

More information

COMP538: Introduction to Bayesian Networks

COMP538: Introduction to Bayesian Networks COMP538: Introduction to ayesian Networks Lecture 4: Inference in ayesian Networks: The VE lgorithm Nevin L. Zhang lzhang@cse.ust.hk Department of Computer Science and Engineering Hong Kong University

More information

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, Networks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Bayesian networks: approximate inference

Bayesian networks: approximate inference Bayesian networks: approximate inference Machine Intelligence Thomas D. Nielsen September 2008 Approximative inference September 2008 1 / 25 Motivation Because of the (worst-case) intractability of exact

More information

Bayesian Network. Outline. Bayesian Network. Syntax Semantics Exact inference by enumeration Exact inference by variable elimination

Bayesian Network. Outline. Bayesian Network. Syntax Semantics Exact inference by enumeration Exact inference by variable elimination Outline Syntax Semantics Exact inference by enumeration Exact inference by variable elimination s A simple, graphical notation for conditional independence assertions and hence for compact specication

More information

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 Plan for today and next week Today and next time: Bayesian networks (Bishop Sec. 8.1) Conditional

More information

CS Lecture 4. Markov Random Fields

CS Lecture 4. Markov Random Fields CS 6347 Lecture 4 Markov Random Fields Recap Announcements First homework is available on elearning Reminder: Office hours Tuesday from 10am-11am Last Time Bayesian networks Today Markov random fields

More information

Bayesian Networks: Belief Propagation in Singly Connected Networks

Bayesian Networks: Belief Propagation in Singly Connected Networks Bayesian Networks: Belief Propagation in Singly Connected Networks Huizhen u janey.yu@cs.helsinki.fi Dept. Computer Science, Univ. of Helsinki Probabilistic Models, Spring, 2010 Huizhen u (U.H.) Bayesian

More information

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 18 The Junction Tree Algorithm Collect & Distribute Algorithmic Complexity ArgMax Junction Tree Algorithm Review: Junction Tree Algorithm end message

More information

Probabilistic Graphical Models: Representation and Inference

Probabilistic Graphical Models: Representation and Inference Probabilistic Graphical Models: Representation and Inference Aaron C. Courville Université de Montréal Note: Material for the slides is taken directly from a presentation prepared by Andrew Moore 1 Overview

More information

Uncertainty and knowledge. Uncertainty and knowledge. Reasoning with uncertainty. Notes

Uncertainty and knowledge. Uncertainty and knowledge. Reasoning with uncertainty. Notes Approximate reasoning Uncertainty and knowledge Introduction All knowledge representation formalism and problem solving mechanisms that we have seen until now are based on the following assumptions: All

More information

Directed Graphical Models or Bayesian Networks

Directed Graphical Models or Bayesian Networks Directed Graphical Models or Bayesian Networks Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Bayesian Networks One of the most exciting recent advancements in statistical AI Compact

More information

Undirected Graphical Models

Undirected Graphical Models Readings: K&F 4. 4.2 4.3 4.4 Undirected Graphical Models Lecture 4 pr 6 20 SE 55 Statistical Methods Spring 20 Instructor: Su-In Lee University of Washington Seattle ayesian Network Representation irected

More information

Bayesian Networks: Independencies and Inference

Bayesian Networks: Independencies and Inference Bayesian Networks: Independencies and Inference Scott Davies and Andrew Moore Note to other teachers and users of these slides. Andrew and Scott would be delighted if you found this source material useful

More information

Probabilistic Graphical Models

Probabilistic Graphical Models 2016 Robert Nowak Probabilistic Graphical Models 1 Introduction We have focused mainly on linear models for signals, in particular the subspace model x = Uθ, where U is a n k matrix and θ R k is a vector

More information

Bayes-Ball: The Rational Pastime (for Determining Irrelevance and Requisite Information in Belief Networks and Influence Diagrams)

Bayes-Ball: The Rational Pastime (for Determining Irrelevance and Requisite Information in Belief Networks and Influence Diagrams) Bayes-Ball: The Rational Pastime (for Determining Irrelevance and Requisite Information in Belief Networks and Influence Diagrams) Ross D. Shachter Engineering-Economic Systems and Operations Research

More information

2. Belief Diagrams. predecessors, C(j), C(j)c:N, and are indicated in

2. Belief Diagrams. predecessors, C(j), C(j)c:N, and are indicated in EVIDENCE ABSORPTION AND PROPAGATION THROUGH EVIDENCE REVERSALS Ross D. Shachter Department of Engineering-Economic Systems Stanford University Stanford, CA 94305-4025 shachter@sumex-aim.stanford.edu The

More information

Learning in Bayesian Networks

Learning in Bayesian Networks Learning in Bayesian Networks Florian Markowetz Max-Planck-Institute for Molecular Genetics Computational Molecular Biology Berlin Berlin: 20.06.2002 1 Overview 1. Bayesian Networks Stochastic Networks

More information

Graphical Models. Andrea Passerini Statistical relational learning. Graphical Models

Graphical Models. Andrea Passerini Statistical relational learning. Graphical Models Andrea Passerini passerini@disi.unitn.it Statistical relational learning Probability distributions Bernoulli distribution Two possible values (outcomes): 1 (success), 0 (failure). Parameters: p probability

More information

Introduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013

Introduction to Bayes Nets. CS 486/686: Introduction to Artificial Intelligence Fall 2013 Introduction to Bayes Nets CS 486/686: Introduction to Artificial Intelligence Fall 2013 1 Introduction Review probabilistic inference, independence and conditional independence Bayesian Networks - - What

More information

Lecture 15. Probabilistic Models on Graph

Lecture 15. Probabilistic Models on Graph Lecture 15. Probabilistic Models on Graph Prof. Alan Yuille Spring 2014 1 Introduction We discuss how to define probabilistic models that use richly structured probability distributions and describe how

More information

UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics. EECS 281A / STAT 241A Statistical Learning Theory

UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics. EECS 281A / STAT 241A Statistical Learning Theory UC Berkeley Department of Electrical Engineering and Computer Science Department of Statistics EECS 281A / STAT 241A Statistical Learning Theory Solutions to Problem Set 2 Fall 2011 Issued: Wednesday,

More information

Inference in Bayesian networks

Inference in Bayesian networks Inference in ayesian networks Devika uramanian omp 440 Lecture 7 xact inference in ayesian networks Inference y enumeration The variale elimination algorithm c Devika uramanian 2006 2 1 roailistic inference

More information

Bayesian & Markov Networks: A unified view

Bayesian & Markov Networks: A unified view School of omputer Science ayesian & Markov Networks: unified view Probabilistic Graphical Models (10-708) Lecture 3, Sep 19, 2007 Receptor Kinase Gene G Receptor X 1 X 2 Kinase Kinase E X 3 X 4 X 5 TF

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Example: multivariate Gaussian Distribution

Example: multivariate Gaussian Distribution School of omputer Science Probabilistic Graphical Models Representation of undirected GM (continued) Eric Xing Lecture 3, September 16, 2009 Reading: KF-chap4 Eric Xing @ MU, 2005-2009 1 Example: multivariate

More information

Applying Bayesian networks in the game of Minesweeper

Applying Bayesian networks in the game of Minesweeper Applying Bayesian networks in the game of Minesweeper Marta Vomlelová Faculty of Mathematics and Physics Charles University in Prague http://kti.mff.cuni.cz/~marta/ Jiří Vomlel Institute of Information

More information

Graphical Model Inference with Perfect Graphs

Graphical Model Inference with Perfect Graphs Graphical Model Inference with Perfect Graphs Tony Jebara Columbia University July 25, 2013 joint work with Adrian Weller Graphical models and Markov random fields We depict a graphical model G as a bipartite

More information

Inference in Bayesian networks

Inference in Bayesian networks Inference in Bayesian networks AIMA2e hapter 14.4 5 1 Outline Exact inference by enumeration Exact inference by variable elimination Approximate inference by stochastic simulation Approximate inference

More information

Markov Independence (Continued)

Markov Independence (Continued) Markov Independence (Continued) As an Undirected Graph Basic idea: Each variable V is represented as a vertex in an undirected graph G = (V(G), E(G)), with set of vertices V(G) and set of edges E(G) the

More information

Graphical Models and Kernel Methods

Graphical Models and Kernel Methods Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.

More information

Probability Propagation in Singly Connected Networks

Probability Propagation in Singly Connected Networks Lecture 4 Probability Propagation in Singly Connected Networks Intelligent Data Analysis and Probabilistic Inference Lecture 4 Slide 1 Probability Propagation We will now develop a general probability

More information

Data Mining 2018 Bayesian Networks (1)

Data Mining 2018 Bayesian Networks (1) Data Mining 2018 Bayesian Networks (1) Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Data Mining 1 / 49 Do you like noodles? Do you like noodles? Race Gender Yes No Black Male 10

More information

Context-specific independence Parameter learning: MLE

Context-specific independence Parameter learning: MLE Use hapter 3 of K&F as a reference for SI Reading for parameter learning: hapter 12 of K&F ontext-specific independence Parameter learning: MLE Graphical Models 10708 arlos Guestrin arnegie Mellon University

More information

Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence

Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence General overview Introduction Directed acyclic graphs (DAGs) and conditional independence DAGs and causal effects

More information

Exact Inference: Clique Trees. Sargur Srihari

Exact Inference: Clique Trees. Sargur Srihari Exact Inference: Clique Trees Sargur srihari@cedar.buffalo.edu 1 Topics 1. Overview 2. Variable Elimination and Clique Trees 3. Message Passing: Sum-Product VE in a Clique Tree Clique-Tree Calibration

More information

Y1 Y2 Y3 Y4 Y1 Y2 Y3 Y4 Z1 Z2 Z3 Z4

Y1 Y2 Y3 Y4 Y1 Y2 Y3 Y4 Z1 Z2 Z3 Z4 Inference: Exploiting Local Structure aphne Koller Stanford University CS228 Handout #4 We have seen that N inference exploits the network structure, in particular the conditional independence and the

More information

Graphical Models 359

Graphical Models 359 8 Graphical Models Probabilities play a central role in modern pattern recognition. We have seen in Chapter 1 that probability theory can be expressed in terms of two simple equations corresponding to

More information

Lecture 6: Graphical Models

Lecture 6: Graphical Models Lecture 6: Graphical Models Kai-Wei Chang CS @ Uniersity of Virginia kw@kwchang.net Some slides are adapted from Viek Skirmar s course on Structured Prediction 1 So far We discussed sequence labeling tasks:

More information

SIMULATION APPROACHES TO GENERAL PROBABILISTIC INFERENCE ON BELIEF NETWORKS

SIMULATION APPROACHES TO GENERAL PROBABILISTIC INFERENCE ON BELIEF NETWORKS SIMULATION APPROACHES TO GENERAL PROBABILISTIC INFERENCE ON BELIEF NETWORKS Ross D. Shachter Department of Engineering-Economic Systems, Stanford University Terman Engineering Center, Stanford, CA 94305-4025

More information

CSC 412 (Lecture 4): Undirected Graphical Models

CSC 412 (Lecture 4): Undirected Graphical Models CSC 412 (Lecture 4): Undirected Graphical Models Raquel Urtasun University of Toronto Feb 2, 2016 R Urtasun (UofT) CSC 412 Feb 2, 2016 1 / 37 Today Undirected Graphical Models: Semantics of the graph:

More information

Decision-Theoretic Specification of Credal Networks: A Unified Language for Uncertain Modeling with Sets of Bayesian Networks

Decision-Theoretic Specification of Credal Networks: A Unified Language for Uncertain Modeling with Sets of Bayesian Networks Decision-Theoretic Specification of Credal Networks: A Unified Language for Uncertain Modeling with Sets of Bayesian Networks Alessandro Antonucci Marco Zaffalon IDSIA, Istituto Dalle Molle di Studi sull

More information

Probabilistic Models. Models describe how (a portion of) the world works

Probabilistic Models. Models describe how (a portion of) the world works Probabilistic Models Models describe how (a portion of) the world works Models are always simplifications May not account for every variable May not account for all interactions between variables All models

More information

Introduction to Probabilistic Graphical Models

Introduction to Probabilistic Graphical Models Introduction to Probabilistic Graphical Models Sargur Srihari srihari@cedar.buffalo.edu 1 Topics 1. What are probabilistic graphical models (PGMs) 2. Use of PGMs Engineering and AI 3. Directionality in

More information

Rapid Introduction to Machine Learning/ Deep Learning

Rapid Introduction to Machine Learning/ Deep Learning Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian

More information

Bayesian Networks Introduction to Machine Learning. Matt Gormley Lecture 24 April 9, 2018

Bayesian Networks Introduction to Machine Learning. Matt Gormley Lecture 24 April 9, 2018 10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Bayesian Networks Matt Gormley Lecture 24 April 9, 2018 1 Homework 7: HMMs Reminders

More information

Lecture 8: Bayesian Networks

Lecture 8: Bayesian Networks Lecture 8: Bayesian Networks Bayesian Networks Inference in Bayesian Networks COMP-652 and ECSE 608, Lecture 8 - January 31, 2017 1 Bayes nets P(E) E=1 E=0 0.005 0.995 E B P(B) B=1 B=0 0.01 0.99 E=0 E=1

More information

4 : Exact Inference: Variable Elimination

4 : Exact Inference: Variable Elimination 10-708: Probabilistic Graphical Models 10-708, Spring 2014 4 : Exact Inference: Variable Elimination Lecturer: Eric P. ing Scribes: Soumya Batra, Pradeep Dasigi, Manzil Zaheer 1 Probabilistic Inference

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Stat 521A Lecture 18 1

Stat 521A Lecture 18 1 Stat 521A Lecture 18 1 Outline Cts and discrete variables (14.1) Gaussian networks (14.2) Conditional Gaussian networks (14.3) Non-linear Gaussian networks (14.4) Sampling (14.5) 2 Hybrid networks A hybrid

More information

CS 5522: Artificial Intelligence II

CS 5522: Artificial Intelligence II CS 5522: Artificial Intelligence II Bayes Nets: Independence Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]

More information

Independencies. Undirected Graphical Models 2: Independencies. Independencies (Markov networks) Independencies (Bayesian Networks)

Independencies. Undirected Graphical Models 2: Independencies. Independencies (Markov networks) Independencies (Bayesian Networks) (Bayesian Networks) Undirected Graphical Models 2: Use d-separation to read off independencies in a Bayesian network Takes a bit of effort! 1 2 (Markov networks) Use separation to determine independencies

More information