6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008

Size: px
Start display at page:

Download "6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008"

Transcription

1 MIT OpenCourseWare / Computational Biology: Genomes, Networks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit:

2 6.047/6.878 Computational Biology: Genomes, Networks, Evolution Introduction to Bayesian Networks

3 Overview We have looked at a number of graphical representations of probability distributions DAG example: HMM Undirected graph example: CRF Today we will look at a very general graphical model representation Bayesian Networks One application modeling gene expression Aviv Regev guest lecture an extension of this basic idea

4 Probabilistic Reconstruction Expression data gives us information about what genes tend to be expressed with others In probability terms, information about the joint distribution over gene states X: P(X)=P(X 1,X 2,X 3,X 4,,X m ) Can we model this joint distribution?

5 Bayesian Networks Directed graph encoding joint distribution variables X P(X) = P(X1,X2,X3,,XN) Learning approaches Inference algorithms Captures information about dependency structure of P(X)

6 Example 1 Icy Roads I Icy Roads W Watson Crashes Causal impact H Holmes Crashes Assume we learn that Watson has crashed Given this causal network, one might fear Holmes has crashed too. Why?

7 Example 1 Icy Roads I Icy Roads W Watson Crashes H Holmes Crashes Now imagine we have learned that roads are not icy We would no longer have an increased fear that Holmes has crashed

8 Conditional Independence We Know I I Icy Roads No Info on I W Watson Crashes H Holmes Crashes If we know nothing about I, W and H are dependent If we know I, W and H are conditionally independent

9 Conditional Independency Independence of 2 random variables X Y PXY (, ) = PXPY ( ) ( ) Conditional independence given a third X Y Z P( X, Y Z) = P( X Z) P( Y Z) but PXY (, ) PXPY ( ) ( ) necessarily

10 Example 2 Rain/Sprinkler R Rain S Sprinkler W Watson Wet H Holmes Wet Holmes discovers his house is wet. Could be rain or his sprinkler.

11 Example 2 Rain/Sprinkler R Rain S Sprinkler W Watson Wet H Holmes Wet Now imagine Holmes sees Watsons grass is wet Now we conclude it was probably rain And probably not his sprinkler

12 Explaining Away R Rain S Sprinkler W Watson Wet H Holmes Wet Initially we had two explanations for Holmes wet grass. But once we had more evidence for R, this explained away H and thus no reason for increase in S

13 Conditional Dependence Don t know H R Rain S Sprinkler W Watson Wet H Holmes Wet Know H If we don t know H, R and S are independent But if we know H, R and S are conditionally dependent

14 Graph Semantics Three basic building blocks X Z X Z X Z Y Y Y Serial Diverging Converging Each implies a particular independence relationship

15 Chain/Linear Conditional Independence X Y Z X Z X Y Z X Z Y

16 Diverging Conditional Independence X Y Z X Z X Y Z X Z Y

17 Converging Conditional Dependence - Explaining Away X Y Z X Z X Y Z X Z Y

18 Graph Semantics Three basic building blocks X Z Y A B. Converging

19 D-Separation Three semantics combine in concept of d-separation Definition : Two nodes A and B are d- separated (or blocked) if for every path p between A and B there is a node V such that either 1. The connection is serial or diverging and V is known 2. The connection is converging and V and all of its descendants are unknown If A and B are not d-separated, they are d- connected

20 Equivalence of Networks Two structures are equivalent if they represent same independence relationship - they encode the same space of probability distributions Example X Z X Z X Z Y Y Y Will return to this when we consider causal vs probabilistic networks

21 Bayesian Networks A Bayesian network (BN) for X = {X 1, X 2, X 3,, X n } consists of: A network structure S Directed acyclic graph (DAG) Nodes => random variables X Encodes graph independence sematics Set of probability distributions P Conditional Probability Distribution (CPD) Local distributions for X

22 Example Bayesian Network R Rain S Sprinkler W Watson Wet H Holmes Wet P( X) = PRPS ( ) ( RPW ) ( RSPH, ) ( RSW,, ) = PRPSPW ( ) ( ) ( RPH ) ( SR, )

23 BN Probability Distribution Only need distributions over nodes and their parents n PX ( ) = PX ( X,..., X ) = i= 1 n i= 1 i 1 i 1 PX ( pax ( )) i i

24 BNs are Compact Descriptions of P(X) Independencies allow us to factorize distribution X2 X1 X3 Example - Assume 2 states per node PX ( ) = P(X )P(X X )P(X X,X ) P(X X,X,X )P(X5 X,X,X,X ) = 62 entries X4 X5 n PX ( ) = PX ( pax ( )) = i= 1 i P(X1)P(X2 X1)P(X3 X1) i P(X4 X2)P(X5 X3) = 18 entries

25 Recall from HMM/CRF Lecture Y 1 Y 2 Y 3 Y 4 Y 1 Y 2 Y 3 Y 4 X 1 X 2 X 3 X 4 X 2 Directed Graph Semantics Potential Functions over Cliques (conditioned on X) Markov Random Field P(X,Y) = Factorization all nodes v P(v parents(v)) ( ) ( ) = P Y Y P X Y i i-1 i i P(Y X) = Factorization all nodes v P(v clique(v),x) ( ) = P Y Y,X i i-1

26 CPDs Discrete R S R S H P(H R,S) W Continuous H

27 Bayesian Networks for Inference Observational inference Observe values (evidence on) of a set of nodes, want to predict state of other nodes Exact Inference Junction Tree Algorithm Approximate Inference Variational approaches, Monte Carlo sampling

28 Walking Through an Example R S P 0 (R) W H P 0 (S) S=y S=n P 0 (W R) W=y W=n P 0 (H R,S) S=y 1,0 0.9,0.1 S=n 0,0 0,1

29 Walking Through an Example R S We define two clusters: -WR, RHS W H The key idea: the clusters only communicate through R If they agree on R, all is good WR R RHS

30 Walking Through an Example We will find it easier to work on this representation: WR R RHS We then need P(WR) and P(RHS): P(WR) =P(R)P(W R) P(RHS) =P(R)P(S)P(H R,S) P 0 (R) P 0 (S) S=y S=n P 0 (W R) W=y W=n P 0 (H R,S) S=y 1,0 0.9,0.1 S=n 0,0 0,1

31 Walking Through an Example We will find it easier to work on this representation: P 0 (R) WR R RHS We then need P(WR) and P(RHS): P(WR) =P(R)P(W R) P(RHS) =P(R)P(S)P(H R,S) P 0 (W,R) W=y W=n P 0 (R,H,S) S=y 0.02, ,0.008 S=n 0.18,0 0,0.72

32 Walking Through an Example P 0 (R) R=y R=n WR R RHS Note that by marginalizing out W from P 0 (W,R) we get P 0 (W)=(0.36,0.64) This is our initial belief in Watsons grass being (wet,not wet) P 0 (W,R) W=y W=n P 0 (R,H,S) S=y 0.02, ,0.008 S=n 0.18,0 0,0.72

33 Walking Through an Example H=y P 0 (R) WR R RHS Now we observe H=y We need to do three things: 1. Update RHS with this info 2. Calculate a new P 1 (R) 3. Transmit P 1 (R) to update WR P 0 (W,R) W=y W=n P 0 (R,H,S) S=y 0.02, ,0.008 S=n 0.18,0 0,0.72

34 Walking Through an Example H=y P 0 (R) WR R RHS Updating RHS with H=y We can simply - Zero out all entries in RHS where H=n P 0 (W,R) W=y W=n P 0 (R,H,S) S=y 0.02, ,0 S=n 0.18,0 0,0

35 Walking Through an Example H=y P 0 (R) WR R RHS Updating RHS with H=y We can simply - Zero out all entries in RHS where H=n - Re-normalize But you can see that this changes P(R) from the perspective of RHS P 0 (W,R) W=y W=n P 1 (R,H,S) S=y 0.074, ,0 S=n 0.662,0 0,0

36 Walking Through an Example H=y P 0 (R) WR R RHS Calculate new P 1 (R) Marginalize out H,S from RHS for: P 1 (R) = (0.736,0.264) Note also P 1 (S) =(0.339,0.661) P 0 (S) =(0.1,0.9) P 0 (W,R) W=y W=n P 1 (R,H,S) S=y 0.074, ,0 S=n 0.662,0 0,0

37 Walking Through an Example P 1 (R) H=y P 1 (R) WR R RHS Transmit P 1 (R) to update WR P (W,R)=P(W R)P (R) =P 0(W,R) P 0 P(R) (R) P 0 (W,R) W=y W=n P 1 (R,H,S) S=y 0.074, ,0 S=n 0.662,0 0,0

38 Walking Through an Example P 1 (R) H=y P 1 (R) WR R RHS Transmit P 1 (R) to update WR P (W,R)=P(W R)P (R) =P 0(W,R) P 0 P(R) (R) P 1 (W,R) W=y W=n P 1 (R,H,S) S=y 0.074, ,0 S=n 0.662,0 0,0

39 Walking Through an Example P 1 (R)/P 0 (R) P 1 (R) H=y P 1 (R) WR R RHS Transmit P 1 (R) to update WR P (W,R)=P(W R)P (R) =P 0(W,R) P 0 P(R) (R) P 1 (W,R) W=y W=n P 1 (R,H,S) S=y 0.074, ,0 S=n 0.662,0 0,0

40 Walking Through an Example P 1 (R)/P 0 (R) P 1 (R) H=y P 1 (R) WR R RHS Transmit P 1 (R) to update WR P (W,R)=P(W R)P (R) 1 1 P 1 0 P(R) 1 =P 0(W,R) P(R) (W=y)=0.788 P(W=y)= P 1 (W,R) W=y W=n P 1 (R,H,S) S=y 0.074, ,0 S=n 0.662,0 0,0

41 Walking Through an Example W=y P 2 (R) P 2 (R)/P 1 (R) H=y P 1 (R) WR R RHS Now we observe W=y 1. Update WR with this info 2. Calculate a new P 2 (R) 3. Transmit P 2 (R) to update WR P 1 (W,R) W=y W=n P 1 (R,H,S) S=y 0.074, ,0 S=n 0.662,0 0,0

42 Walking Through an Example W=y P 2 (R) P 2 (R)/P 1 (R) H=y P 2 (R) WR R RHS P 2 (S=y)=0.161 P 1 (S=y)=0.339 P 0 (S=y)=0.1 - R is almost certain - We have explained away H=y -S goes low again P 2 (W,R) W=y W=n 0 0 P 2 (R,H,S) S=y 0.094, ,0 S=n 0.839,0 0,0

43 Message Passing in Junction Trees P*(S)/P(S) e X S Y Separator State of separator S is information shared between X and Y When Y is updated, it sends a message to X Message has information to update X to agree with Y on state of S

44 Message Passing in Junction Trees 2 G 3 4 H 1 e e F E D 1 1 A B C A node can send one message to a neighbor, only after receiving all messages from each other neighbor When messages have been passed both ways along a link, it is consistent Passing continues until all links are consistent

45 HUGIN Algorithm A simple algorithm for coordinated message passing in junction trees Select one node, V, as root Call CollectEvidence(V): Ask all neighbors of V to send message to V. If not possible, recursively pass message to all neighbors but V Call DistributeEvidence(V): Send message to all neighbors of V Recursively send message to all neighbors but V

46 BN to Junction Tree R S A topic by itself. In summary: W H Moral Graph undirected graph with links between all parents of all nodes Triangulate add links so all cycles>3 have cord Cliques become nodes of Junction Tree WR R RHS

47 Recall from HMM/CRF Lecture Y 1 Y 2 Y 3 Y 4 Y 1 Y 2 Y 3 Y 4 X 1 X 2 X 3 X 4 X 2 Directed Graph Semantics Potential Functions over Cliques (conditioned on X) Markov Random Field P(X,Y) = Factorization all nodes v P(v parents(v)) ( ) ( ) = P Y Y P X Y i i-1 i i P(Y X) = Factorization all nodes v P(v clique(v),x) ( ) = P Y Y,X i i-1

48 Applications G9 G8 G10 G7 G6 G1 G2 G4 G3 G5 Measure some gene expression predict rest

49 Latent Variables G9 G8 Drugs D1 G10 G7 G6 D2 G1 G2 G4 G3 G5 D3 D4

50 Latent Variables Env Drugs ph G9 G8 D1 O 2 G10 G7 G6 D2 G1 G2 G4 G3 G5 D3 D4

51 Latent Variables Env Drugs ph G9 G8 D1 O 2 G10 G7 G6 D2 G1 G2 G4 G3 G5 D3 D4 Metabolites M1 M2 M3 M4

52 Observation vs Intervention Arrows not necessarily causal BN models probability and correlation between variables For applications so far, we observe evidence and want to know states of other nodes most likely to go with observation What about interventions?

53 Example Sprinkler W Season Rain H Sprinkler If we observe the Sprinkler on Holmes grass more likely wet And more likely summer But what if we force sprinkler on Intervention cut arrows from parents

54 Causal vs Probabilistic X Y Z This depends on getting the arrows correct! X Y Z Flipping all arrows does not change independence relationships But changes causality for interventions

55 Learning Bayesian Networks Given a set of observations D (i.e. expression data set) on X, we want to find: 1. A network structure S 2. Parameters, Θ, for probability distributions on each node, given S Relatively Easy

56 Learning Θ Given S, we can choose maximum likelihood parameter Θ n θ arg max PD ( θ, S) PX ( pax ( ), θ) = = θ i= 1 i i We can also choose to include prior information P(Θ) in a bayesian approach P( θ S,D) = P( S, D θ) P( θ) θbayes = θp( θ S, D) dθ

57 Learning Bayesian Networks Given a set of observations D (i.e. expression data set) on X, we want to find: 1. A network structure S NP-Hard 2. Parameters, Θ, for probability distributions on each node, given S

58 Learning S Find optimal structure S given D PS ( D) PD ( SPS ) ( ) P( D S) = P( D, S) P( S) d θ θ θ In special circumstances, integral analytically tractable (e.g. no missing data, multinomial, dirichlet priors)

59 Learning S Heuristic Search Compute P(S D) for all networks S Select S* that maximizes P(S D) Problem 1: number of S grows super-exponentially with number of nodes no exhaustive search, use hill climbing, etc.. Problem 2: Sparse data and overfitting P(S D) P(S D) S* S S* S Lots of data Sparse data

60 Model Averaging Rather than select only one model as result of search, we can draw many samples from P(M D) and model average PE ( D) = PExy ( DSPS, ) ( D) xy = samples samples 1 ( SPS ) ( D) xy How do we sample...?

61 Sampling Models - MCMC Sample from Markov Chain Monte Carlo Method Direct approach intractable due to partition function MCMC PD ( Ss) PS ( k) PS ( D) = P( D Ss ) P( Sk ) Propose Given S old, propose new S new with probability Q(S new S old ) Accept/Reject Accept S new as sample with PD ( Snew) PS ( new) QS ( old S ) new p = min 1, PD ( Sold ) PS ( old ) QS ( new Sold ) k

Introduction to Bayesian Networks

Introduction to Bayesian Networks Introduction to Bayesian Networks Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1/23 Outline Basic Concepts Bayesian

More information

Statistical Approaches to Learning and Discovery

Statistical Approaches to Learning and Discovery Statistical Approaches to Learning and Discovery Graphical Models Zoubin Ghahramani & Teddy Seidenfeld zoubin@cs.cmu.edu & teddy@stat.cmu.edu CALD / CS / Statistics / Philosophy Carnegie Mellon University

More information

Probabilistic Graphical Networks: Definitions and Basic Results

Probabilistic Graphical Networks: Definitions and Basic Results This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical

More information

9 Forward-backward algorithm, sum-product on factor graphs

9 Forward-backward algorithm, sum-product on factor graphs Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous

More information

An Introduction to Bayesian Machine Learning

An Introduction to Bayesian Machine Learning 1 An Introduction to Bayesian Machine Learning José Miguel Hernández-Lobato Department of Engineering, Cambridge University April 8, 2013 2 What is Machine Learning? The design of computational systems

More information

Graphical Models and Kernel Methods

Graphical Models and Kernel Methods Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.

More information

A Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004

A Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004 A Brief Introduction to Graphical Models Presenter: Yijuan Lu November 12,2004 References Introduction to Graphical Models, Kevin Murphy, Technical Report, May 2001 Learning in Graphical Models, Michael

More information

Learning in Bayesian Networks

Learning in Bayesian Networks Learning in Bayesian Networks Florian Markowetz Max-Planck-Institute for Molecular Genetics Computational Molecular Biology Berlin Berlin: 20.06.2002 1 Overview 1. Bayesian Networks Stochastic Networks

More information

Probabilistic Graphical Models (I)

Probabilistic Graphical Models (I) Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random

More information

Machine Learning Summer School

Machine Learning Summer School Machine Learning Summer School Lecture 3: Learning parameters and structure Zoubin Ghahramani zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin/ Department of Engineering University of Cambridge,

More information

COMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma

COMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma COMS 4771 Probabilistic Reasoning via Graphical Models Nakul Verma Last time Dimensionality Reduction Linear vs non-linear Dimensionality Reduction Principal Component Analysis (PCA) Non-linear methods

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft

More information

An Introduction to Reversible Jump MCMC for Bayesian Networks, with Application

An Introduction to Reversible Jump MCMC for Bayesian Networks, with Application An Introduction to Reversible Jump MCMC for Bayesian Networks, with Application, CleverSet, Inc. STARMAP/DAMARS Conference Page 1 The research described in this presentation has been funded by the U.S.

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: Tony Jebara Topic 16 Undirected Graphs Undirected Separation Inferring Marginals & Conditionals Moralization Junction Trees Triangulation Undirected Graphs Separation

More information

6.867 Machine learning, lecture 23 (Jaakkola)

6.867 Machine learning, lecture 23 (Jaakkola) Lecture topics: Markov Random Fields Probabilistic inference Markov Random Fields We will briefly go over undirected graphical models or Markov Random Fields (MRFs) as they will be needed in the context

More information

Lecture 8: Bayesian Networks

Lecture 8: Bayesian Networks Lecture 8: Bayesian Networks Bayesian Networks Inference in Bayesian Networks COMP-652 and ECSE 608, Lecture 8 - January 31, 2017 1 Bayes nets P(E) E=1 E=0 0.005 0.995 E B P(B) B=1 B=0 0.01 0.99 E=0 E=1

More information

Bayesian networks. Instructor: Vincent Conitzer

Bayesian networks. Instructor: Vincent Conitzer Bayesian networks Instructor: Vincent Conitzer Rain and sprinklers example raining (X) sprinklers (Y) P(X=1) =.3 P(Y=1) =.4 grass wet (Z) Each node has a conditional probability table (CPT) P(Z=1 X=0,

More information

Chapter 16. Structured Probabilistic Models for Deep Learning

Chapter 16. Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 1 Chapter 16 Structured Probabilistic Models for Deep Learning Peng et al.: Deep Learning and Practice 2 Structured Probabilistic Models way of using graphs to describe

More information

2 : Directed GMs: Bayesian Networks

2 : Directed GMs: Bayesian Networks 10-708: Probabilistic Graphical Models 10-708, Spring 2017 2 : Directed GMs: Bayesian Networks Lecturer: Eric P. Xing Scribes: Jayanth Koushik, Hiroaki Hayashi, Christian Perez Topic: Directed GMs 1 Types

More information

The Monte Carlo Method: Bayesian Networks

The Monte Carlo Method: Bayesian Networks The Method: Bayesian Networks Dieter W. Heermann Methods 2009 Dieter W. Heermann ( Methods)The Method: Bayesian Networks 2009 1 / 18 Outline 1 Bayesian Networks 2 Gene Expression Data 3 Bayesian Networks

More information

Lecture 6: Graphical Models: Learning

Lecture 6: Graphical Models: Learning Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)

More information

Outline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination

Outline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination Probabilistic Graphical Models COMP 790-90 Seminar Spring 2011 The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Outline It Introduction ti Representation Bayesian network Conditional Independence Inference:

More information

Inference in Graphical Models Variable Elimination and Message Passing Algorithm

Inference in Graphical Models Variable Elimination and Message Passing Algorithm Inference in Graphical Models Variable Elimination and Message Passing lgorithm Le Song Machine Learning II: dvanced Topics SE 8803ML, Spring 2012 onditional Independence ssumptions Local Markov ssumption

More information

p L yi z n m x N n xi

p L yi z n m x N n xi y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen

More information

CS 188: Artificial Intelligence. Bayes Nets

CS 188: Artificial Intelligence. Bayes Nets CS 188: Artificial Intelligence Probabilistic Inference: Enumeration, Variable Elimination, Sampling Pieter Abbeel UC Berkeley Many slides over this course adapted from Dan Klein, Stuart Russell, Andrew

More information

CS281A/Stat241A Lecture 19

CS281A/Stat241A Lecture 19 CS281A/Stat241A Lecture 19 p. 1/4 CS281A/Stat241A Lecture 19 Junction Tree Algorithm Peter Bartlett CS281A/Stat241A Lecture 19 p. 2/4 Announcements My office hours: Tuesday Nov 3 (today), 1-2pm, in 723

More information

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)

More information

CS Lecture 3. More Bayesian Networks

CS Lecture 3. More Bayesian Networks CS 6347 Lecture 3 More Bayesian Networks Recap Last time: Complexity challenges Representing distributions Computing probabilities/doing inference Introduction to Bayesian networks Today: D-separation,

More information

Directed and Undirected Graphical Models

Directed and Undirected Graphical Models Directed and Undirected Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Last Lecture Refresher Lecture Plan Directed

More information

CSC 412 (Lecture 4): Undirected Graphical Models

CSC 412 (Lecture 4): Undirected Graphical Models CSC 412 (Lecture 4): Undirected Graphical Models Raquel Urtasun University of Toronto Feb 2, 2016 R Urtasun (UofT) CSC 412 Feb 2, 2016 1 / 37 Today Undirected Graphical Models: Semantics of the graph:

More information

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic

Announcements. CS 188: Artificial Intelligence Fall Causality? Example: Traffic. Topology Limits Distributions. Example: Reverse Traffic CS 188: Artificial Intelligence Fall 2008 Lecture 16: Bayes Nets III 10/23/2008 Announcements Midterms graded, up on glookup, back Tuesday W4 also graded, back in sections / box Past homeworks in return

More information

Probabilistic Graphical Models. Guest Lecture by Narges Razavian Machine Learning Class April

Probabilistic Graphical Models. Guest Lecture by Narges Razavian Machine Learning Class April Probabilistic Graphical Models Guest Lecture by Narges Razavian Machine Learning Class April 14 2017 Today What is probabilistic graphical model and why it is useful? Bayesian Networks Basic Inference

More information

CS Lecture 4. Markov Random Fields

CS Lecture 4. Markov Random Fields CS 6347 Lecture 4 Markov Random Fields Recap Announcements First homework is available on elearning Reminder: Office hours Tuesday from 10am-11am Last Time Bayesian networks Today Markov random fields

More information

Markov Networks.

Markov Networks. Markov Networks www.biostat.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts Markov network syntax Markov network semantics Potential functions Partition function

More information

Undirected Graphical Models: Markov Random Fields

Undirected Graphical Models: Markov Random Fields Undirected Graphical Models: Markov Random Fields 40-956 Advanced Topics in AI: Probabilistic Graphical Models Sharif University of Technology Soleymani Spring 2015 Markov Random Field Structure: undirected

More information

Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo. Sampling Methods. Machine Learning. Torsten Möller.

Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo. Sampling Methods. Machine Learning. Torsten Möller. Sampling Methods Machine Learning orsten Möller Möller/Mori 1 Recall Inference or General Graphs Junction tree algorithm is an exact inference method for arbitrary graphs A particular tree structure defined

More information

Implementing Machine Reasoning using Bayesian Network in Big Data Analytics

Implementing Machine Reasoning using Bayesian Network in Big Data Analytics Implementing Machine Reasoning using Bayesian Network in Big Data Analytics Steve Cheng, Ph.D. Guest Speaker for EECS 6893 Big Data Analytics Columbia University October 26, 2017 Outline Introduction Probability

More information

Rapid Introduction to Machine Learning/ Deep Learning

Rapid Introduction to Machine Learning/ Deep Learning Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian

More information

Based on slides by Richard Zemel

Based on slides by Richard Zemel CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we

More information

Tópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863

Tópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863 Tópicos Especiais em Modelagem e Análise - Aprendizado por Máquina CPS863 Daniel, Edmundo, Rosa Terceiro trimestre de 2012 UFRJ - COPPE Programa de Engenharia de Sistemas e Computação Bayesian Networks

More information

Summary of the Bayes Net Formalism. David Danks Institute for Human & Machine Cognition

Summary of the Bayes Net Formalism. David Danks Institute for Human & Machine Cognition Summary of the Bayes Net Formalism David Danks Institute for Human & Machine Cognition Bayesian Networks Two components: 1. Directed Acyclic Graph (DAG) G: There is a node for every variable D: Some nodes

More information

The Origin of Deep Learning. Lili Mou Jan, 2015

The Origin of Deep Learning. Lili Mou Jan, 2015 The Origin of Deep Learning Lili Mou Jan, 2015 Acknowledgment Most of the materials come from G. E. Hinton s online course. Outline Introduction Preliminary Boltzmann Machines and RBMs Deep Belief Nets

More information

Probabilistic Models

Probabilistic Models Bayes Nets 1 Probabilistic Models Models describe how (a portion of) the world works Models are always simplifications May not account for every variable May not account for all interactions between variables

More information

Bayesian Networks Inference with Probabilistic Graphical Models

Bayesian Networks Inference with Probabilistic Graphical Models 4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning

More information

Machine Learning Lecture 14

Machine Learning Lecture 14 Many slides adapted from B. Schiele, S. Roth, Z. Gharahmani Machine Learning Lecture 14 Undirected Graphical Models & Inference 23.06.2015 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de/ leibe@vision.rwth-aachen.de

More information

Junction Tree, BP and Variational Methods

Junction Tree, BP and Variational Methods Junction Tree, BP and Variational Methods Adrian Weller MLSALT4 Lecture Feb 21, 2018 With thanks to David Sontag (MIT) and Tony Jebara (Columbia) for use of many slides and illustrations For more information,

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should

More information

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Problem Set 3 Issued: Thursday, September 25, 2014 Due: Thursday,

More information

Introduction to Bayesian Learning

Introduction to Bayesian Learning Course Information Introduction Introduction to Bayesian Learning Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Apprendimento Automatico: Fondamenti - A.A. 2016/2017 Outline

More information

Stochastic inference in Bayesian networks, Markov chain Monte Carlo methods

Stochastic inference in Bayesian networks, Markov chain Monte Carlo methods Stochastic inference in Bayesian networks, Markov chain Monte Carlo methods AI: Stochastic inference in BNs AI: Stochastic inference in BNs 1 Outline ypes of inference in (causal) BNs Hardness of exact

More information

4 : Exact Inference: Variable Elimination

4 : Exact Inference: Variable Elimination 10-708: Probabilistic Graphical Models 10-708, Spring 2014 4 : Exact Inference: Variable Elimination Lecturer: Eric P. ing Scribes: Soumya Batra, Pradeep Dasigi, Manzil Zaheer 1 Probabilistic Inference

More information

Part I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS

Part I. C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Part I C. M. Bishop PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 8: GRAPHICAL MODELS Probabilistic Graphical Models Graphical representation of a probabilistic model Each variable corresponds to a

More information

Icy Roads. Bayesian Networks. Icy Roads. Icy Roads. Icy Roads. Holmes and Watson in LA

Icy Roads. Bayesian Networks. Icy Roads. Icy Roads. Icy Roads. Holmes and Watson in LA 6.825 Techniques in rtificial ntelligence ayesian Networks To do probabilistic reasoning, you need to know the joint probability distribution ut, in a domain with N propositional variables, one needs 2

More information

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

A graph contains a set of nodes (vertices) connected by links (edges or arcs) BOLTZMANN MACHINES Generative Models Graphical Models A graph contains a set of nodes (vertices) connected by links (edges or arcs) In a probabilistic graphical model, each node represents a random variable,

More information

Alternative Parameterizations of Markov Networks. Sargur Srihari

Alternative Parameterizations of Markov Networks. Sargur Srihari Alternative Parameterizations of Markov Networks Sargur srihari@cedar.buffalo.edu 1 Topics Three types of parameterization 1. Gibbs Parameterization 2. Factor Graphs 3. Log-linear Models Features (Ising,

More information

Generative and Discriminative Approaches to Graphical Models CMSC Topics in AI

Generative and Discriminative Approaches to Graphical Models CMSC Topics in AI Generative and Discriminative Approaches to Graphical Models CMSC 35900 Topics in AI Lecture 2 Yasemin Altun January 26, 2007 Review of Inference on Graphical Models Elimination algorithm finds single

More information

Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo. Sampling Methods. Oliver Schulte - CMPT 419/726. Bishop PRML Ch.

Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo. Sampling Methods. Oliver Schulte - CMPT 419/726. Bishop PRML Ch. Sampling Methods Oliver Schulte - CMP 419/726 Bishop PRML Ch. 11 Recall Inference or General Graphs Junction tree algorithm is an exact inference method for arbitrary graphs A particular tree structure

More information

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models 02-710 Computational Genomics Systems biology Putting it together: Data integration using graphical models High throughput data So far in this class we discussed several different types of high throughput

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Undirected Graphical Models Mark Schmidt University of British Columbia Winter 2016 Admin Assignment 3: 2 late days to hand it in today, Thursday is final day. Assignment 4:

More information

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation

Announcements. CS 188: Artificial Intelligence Spring Probability recap. Outline. Bayes Nets: Big Picture. Graphical Model Notation CS 188: Artificial Intelligence Spring 2010 Lecture 15: Bayes Nets II Independence 3/9/2010 Pieter Abbeel UC Berkeley Many slides over the course adapted from Dan Klein, Stuart Russell, Andrew Moore Current

More information

Alternative Parameterizations of Markov Networks. Sargur Srihari

Alternative Parameterizations of Markov Networks. Sargur Srihari Alternative Parameterizations of Markov Networks Sargur srihari@cedar.buffalo.edu 1 Topics Three types of parameterization 1. Gibbs Parameterization 2. Factor Graphs 3. Log-linear Models with Energy functions

More information

Chris Bishop s PRML Ch. 8: Graphical Models

Chris Bishop s PRML Ch. 8: Graphical Models Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular

More information

Bayes Nets: Independence

Bayes Nets: Independence Bayes Nets: Independence [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Bayes Nets A Bayes

More information

PROBABILISTIC REASONING SYSTEMS

PROBABILISTIC REASONING SYSTEMS PROBABILISTIC REASONING SYSTEMS In which we explain how to build reasoning systems that use network models to reason with uncertainty according to the laws of probability theory. Outline Knowledge in uncertain

More information

CS 343: Artificial Intelligence

CS 343: Artificial Intelligence CS 343: Artificial Intelligence Bayes Nets: Sampling Prof. Scott Niekum The University of Texas at Austin [These slides based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.

More information

Bayes Networks 6.872/HST.950

Bayes Networks 6.872/HST.950 Bayes Networks 6.872/HST.950 What Probabilistic Models Should We Use? Full joint distribution Completely expressive Hugely data-hungry Exponential computational complexity Naive Bayes (full conditional

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

Lecture 6: Graphical Models

Lecture 6: Graphical Models Lecture 6: Graphical Models Kai-Wei Chang CS @ Uniersity of Virginia kw@kwchang.net Some slides are adapted from Viek Skirmar s course on Structured Prediction 1 So far We discussed sequence labeling tasks:

More information

Intelligent Systems (AI-2)

Intelligent Systems (AI-2) Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 11 Oct, 3, 2016 CPSC 422, Lecture 11 Slide 1 422 big picture: Where are we? Query Planning Deterministic Logics First Order Logics Ontologies

More information

Bayesian Machine Learning

Bayesian Machine Learning Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September

More information

Variational Scoring of Graphical Model Structures

Variational Scoring of Graphical Model Structures Variational Scoring of Graphical Model Structures Matthew J. Beal Work with Zoubin Ghahramani & Carl Rasmussen, Toronto. 15th September 2003 Overview Bayesian model selection Approximations using Variational

More information

Probabilistic Graphical Models: Representation and Inference

Probabilistic Graphical Models: Representation and Inference Probabilistic Graphical Models: Representation and Inference Aaron C. Courville Université de Montréal Note: Material for the slides is taken directly from a presentation prepared by Andrew Moore 1 Overview

More information

Sampling Methods. Bishop PRML Ch. 11. Alireza Ghane. Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo

Sampling Methods. Bishop PRML Ch. 11. Alireza Ghane. Sampling Rejection Sampling Importance Sampling Markov Chain Monte Carlo Sampling Methods Bishop PRML h. 11 Alireza Ghane Sampling Methods A. Ghane /. Möller / G. Mori 1 Recall Inference or General Graphs Junction tree algorithm is an exact inference method for arbitrary graphs

More information

Probabilistic Graphical Models

Probabilistic Graphical Models 2016 Robert Nowak Probabilistic Graphical Models 1 Introduction We have focused mainly on linear models for signals, in particular the subspace model x = Uθ, where U is a n k matrix and θ R k is a vector

More information

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech

Probabilistic Graphical Models and Bayesian Networks. Artificial Intelligence Bert Huang Virginia Tech Probabilistic Graphical Models and Bayesian Networks Artificial Intelligence Bert Huang Virginia Tech Concept Map for Segment Probabilistic Graphical Models Probabilistic Time Series Models Particle Filters

More information

Lecture 17: May 29, 2002

Lecture 17: May 29, 2002 EE596 Pat. Recog. II: Introduction to Graphical Models University of Washington Spring 2000 Dept. of Electrical Engineering Lecture 17: May 29, 2002 Lecturer: Jeff ilmes Scribe: Kurt Partridge, Salvador

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Probabilistic Graphical Models Homework 2: Due February 24, 2014 at 4 pm

Probabilistic Graphical Models Homework 2: Due February 24, 2014 at 4 pm Probabilistic Graphical Models 10-708 Homework 2: Due February 24, 2014 at 4 pm Directions. This homework assignment covers the material presented in Lectures 4-8. You must complete all four problems to

More information

Bayesian Networks. Motivation

Bayesian Networks. Motivation Bayesian Networks Computer Sciences 760 Spring 2014 http://pages.cs.wisc.edu/~dpage/cs760/ Motivation Assume we have five Boolean variables,,,, The joint probability is,,,, How many state configurations

More information

Bayes Nets III: Inference

Bayes Nets III: Inference 1 Hal Daumé III (me@hal3.name) Bayes Nets III: Inference Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 10 Apr 2012 Many slides courtesy

More information

Directed Graphical Models

Directed Graphical Models Directed Graphical Models Instructor: Alan Ritter Many Slides from Tom Mitchell Graphical Models Key Idea: Conditional independence assumptions useful but Naïve Bayes is extreme! Graphical models express

More information

Introduction to Bayesian inference

Introduction to Bayesian inference Introduction to Bayesian inference Thomas Alexander Brouwer University of Cambridge tab43@cam.ac.uk 17 November 2015 Probabilistic models Describe how data was generated using probability distributions

More information

Advanced Probabilistic Modeling in R Day 1

Advanced Probabilistic Modeling in R Day 1 Advanced Probabilistic Modeling in R Day 1 Roger Levy University of California, San Diego July 20, 2015 1/24 Today s content Quick review of probability: axioms, joint & conditional probabilities, Bayes

More information

Bayesian Networks to design optimal experiments. Davide De March

Bayesian Networks to design optimal experiments. Davide De March Bayesian Networks to design optimal experiments Davide De March davidedemarch@gmail.com 1 Outline evolutionary experimental design in high-dimensional space and costly experimentation the microwell mixture

More information

T Machine Learning: Basic Principles

T Machine Learning: Basic Principles Machine Learning: Basic Principles Bayesian Networks Laboratory of Computer and Information Science (CIS) Department of Computer Science and Engineering Helsinki University of Technology (TKK) Autumn 2007

More information

Lecture 15. Probabilistic Models on Graph

Lecture 15. Probabilistic Models on Graph Lecture 15. Probabilistic Models on Graph Prof. Alan Yuille Spring 2014 1 Introduction We discuss how to define probabilistic models that use richly structured probability distributions and describe how

More information

Sampling Algorithms for Probabilistic Graphical models

Sampling Algorithms for Probabilistic Graphical models Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Matrix Data: Classification: Part 2 Instructor: Yizhou Sun yzsun@ccs.neu.edu September 21, 2014 Methods to Learn Matrix Data Set Data Sequence Data Time Series Graph & Network

More information

Directed Graphical Models or Bayesian Networks

Directed Graphical Models or Bayesian Networks Directed Graphical Models or Bayesian Networks Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Bayesian Networks One of the most exciting recent advancements in statistical AI Compact

More information

Soft Computing. Lecture Notes on Machine Learning. Matteo Matteucci.

Soft Computing. Lecture Notes on Machine Learning. Matteo Matteucci. Soft Computing Lecture Notes on Machine Learning Matteo Matteucci matteucci@elet.polimi.it Department of Electronics and Information Politecnico di Milano Matteo Matteucci c Lecture Notes on Machine Learning

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Probabilistic Machine Learning

Probabilistic Machine Learning Probabilistic Machine Learning Bayesian Nets, MCMC, and more Marek Petrik 4/18/2017 Based on: P. Murphy, K. (2012). Machine Learning: A Probabilistic Perspective. Chapter 10. Conditional Independence Independent

More information

1 Undirected Graphical Models. 2 Markov Random Fields (MRFs)

1 Undirected Graphical Models. 2 Markov Random Fields (MRFs) Machine Learning (ML, F16) Lecture#07 (Thursday Nov. 3rd) Lecturer: Byron Boots Undirected Graphical Models 1 Undirected Graphical Models In the previous lecture, we discussed directed graphical models.

More information

CS 5522: Artificial Intelligence II

CS 5522: Artificial Intelligence II CS 5522: Artificial Intelligence II Bayes Nets: Independence Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]

More information

Probabilistic Graphical Models for Image Analysis - Lecture 1

Probabilistic Graphical Models for Image Analysis - Lecture 1 Probabilistic Graphical Models for Image Analysis - Lecture 1 Alexey Gronskiy, Stefan Bauer 21 September 2018 Max Planck ETH Center for Learning Systems Overview 1. Motivation - Why Graphical Models 2.

More information

13: Variational inference II

13: Variational inference II 10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational

More information

Learning With Bayesian Networks. Markus Kalisch ETH Zürich

Learning With Bayesian Networks. Markus Kalisch ETH Zürich Learning With Bayesian Networks Markus Kalisch ETH Zürich Inference in BNs - Review P(Burglary JohnCalls=TRUE, MaryCalls=TRUE) Exact Inference: P(b j,m) = c Sum e Sum a P(b)P(e)P(a b,e)p(j a)p(m a) Deal

More information

Bayesian Machine Learning - Lecture 7

Bayesian Machine Learning - Lecture 7 Bayesian Machine Learning - Lecture 7 Guido Sanguinetti Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh gsanguin@inf.ed.ac.uk March 4, 2015 Today s lecture 1

More information

Bayesian Networks Basic and simple graphs

Bayesian Networks Basic and simple graphs Bayesian Networks Basic and simple graphs Ullrika Sahlin, Centre of Environmental and Climate Research Lund University, Sweden Ullrika.Sahlin@cec.lu.se http://www.cec.lu.se/ullrika-sahlin Bayesian [Belief]

More information