Graph structure learning for network inference

Size: px
Start display at page:

Download "Graph structure learning for network inference"

Transcription

1 Graph structure learning for network inference Sushmita Roy Computa9onal Network Biology Biosta2s2cs & Medical Informa2cs 826 Computer Sciences 838 hbps://compnetbiocourse.discovery.wisc.edu Sep 13 th 2016 Some of the material covered in this lecture is adapted from BMI 576

2 Goals for today Represen2ng transcrip2onal networks Learning regulatory networks from gene expression data Per-gene regulatory network inference algorithms Different types of probabilis2c graphical models Bayesian networks

3 Transcrip9onal regulatory networks Transcrip2on factors (TF) (a) DNA A B Gene C A B C Directed, signed, weighted graph Nodes: TFs and Target genes Edges: A regulates C s expression level Regulatory network of E. coli. 153 TFs (green & light red), 1319 targets Vargas and San2llan, 2008

4 Why do we need to computa9onally infer transcrip9onal networks? Why infer transcrip2onal networks? Control which genes are expressed when and where Needed for accurate informa2on processing in cells Many diseases are associated with changes in transcrip2onal networks Why do so computa2onally? Experimental detec2on of networks is hard, expensive A first step towards having an in silico model of a cell A model can be used to make predic2ons that can be tested and refine the model

5 Methods to infer regulatory networks Experimental ChIP-chip and ChIP-seq Factor/regulator knockout followed by genome-wide transcriptome measurements DNase I/ATAC-seq + mo2f finding Computa2onal Supervised network inference Unsupervised network inference Expression-based network inference

6 Types of data for reconstruc9ng transcrip9onal networks Node-specific datasets Transcriptomes: Genome-wide mrna levels Can poten2ally recover genome-wide regulatory networks Gene expression levels Samples Edge-specific datasets ChIP-chip and ChIP-seq Sequence specific mo2fs Factor knockout followed by wholetranscriptome profiling ChIP mo2f Gene

7 Experimental techniques to measure expression Microarrays cdna/spobed arrays Oligonucleo2des arrays Sequencing RNA-seq

8 Microarrays A microarray is a solid support, on which pieces of DNA are arranged in a grid-like array Each piece is called a probe Measures RNA abundances by exploi2ng complementary hybridiza2on DNA from labeled sample is called target

9 SpoEed versus oligonucleo9de array Three key steps: Reverse transcrip2on, labeling and hybridiza2on From Vermeeren and Luc Michiels 2011 DOI: /19432

10 A video about two color DNA microarrays From: hbps://

11 A video about DNA microarrays From: hbps:// Also see: hbp://

12 PERSP ECTIVES A typical RNA-seq pipeline AAAAAAAA mrna RNA fragments or AAAAAAAA TTTTTTTT cdna EST library with adaptors ATCACAGTGGGACTCCATAAATTTTTCT CGAAGGACCAGCAGAAACGAGAGAAAAA GGACAGAGTCCCCAGCGGGCTGAAGGGG ATGAAACATTAAAGTCAAACAATATGAA... Short sequence reads ORF Coding sequence Junction reads Exonic reads...aaaaaaaaa...aaaaaa poly(a) end reads Mapped sequence reads Wang et al, Nature Gene2cs 2009; RNA expression level Base-resolution expression profile Nucleotide position Figure 1 A typical RNA-Seq experiment. Briefly, long RNAs are first converted into a library of cdna fragments through either RNA fragmentation or DNA fragmentation (see main text). Sequencing adaptors (blue) are subsequently added to each cdna fragment and a short sequence is obtained from

13 What do we want a model for a regulatory network to capture? Hot1 Sko1 HSP12 ON Sko1 HSP12 OFF HSP12 HSP12 HSP12 s expression is dependent upon Hot1 and Sko1 binding to HSP12 s promoter Hot1 Sko1 X 1 X 2 Hot1 regulates HSP12 HSP12 is a target of Hot1 HSP12 X 3 =ψ(x 1,X 2 ) X 3 BOOLEAN LINEAR DIFF. EQNS PROBABILISTIC. Structure Who are the regulators? Func9on How they determine expression levels?

14 Mathema9cal representa9ons of the how ques9on X 1 X 2 Input expression of neighbors ψ Models differ in the func9on that maps input system state to output state X 3 Output expression of target Boolean Networks Differen2al equa2ons Probabilis2c graphical models X1 X2 X3 Input Output X1 X2 X dx 3 (t) = dt g(x 1 (t),x 2 (t)) rx 3 (t) P (X 3 X 1,X 2 )= N(X 1 a + X 2 b, ) Probability distribu9ons

15 Expression-based network inference X 1 X 2 Biological samples Phenotypic Variation in Yeast Genes Expression-based network inference Expression level of gene i in experiment j Figure 3. Variation in gene expression in S. cerevisiae isolates. The diagrams show the average log2 expression differences measured in the denoted strains. Each row represents a given gene and each column represents a different strain, color-coded as described in Figure 1. (A) Expression patterns of 2,680 genes that varied significantly (FDR = 0.01, paired t-test) in at least one strain compared to S288c. (B) Expression patterns of 953 genes that varied significantly in at least one strain compared to strain YPS163 (FDR=0.01, unpaired t-test). For (A) and (B), a red color indicates higher expression and a green color represents lower expression in the denoted strain compared to S288c, according to the key. (C) Expression patterns of 1,330 genes that varied significantly (FDR = 0.01, paired t-test) in at least one strain compared to the mean expression of all 17 strains. Here, red and green correspond to higher and lower expression, respectively, compared to the mean expression of that gene in all strains. Genes were organized independently in each plot by hierarchical clustering. doi: /journal.pgen g003 differences from the mean ranged from 30 (in vineyard strain I14) to nearly 600 (in clinical isolate YJM789), with a median of 88 expression differences per strain. The number of expression differences did not correlate strongly with the genetic distances of the strains (R 2 = 0.16). However, this is not surprising since many Structure sensory protein [35]. We also noted that the 1330 genes with statistically variable expression in at least one non-laboratory of the observed expression differences are likely linked in trans to the same genetic loci [27,31,34,35,43]. Consistent with this of some of these categories was also observed for a different set of vineyard strains [26,28], and the genetic basis for differential expression of amino acid biosynthetic genes in one vineyard strain has recently been linked to a polymorphism in an amino acid strain were enriched for genes that contained upstream TATA X 3 X 1 X 2 X 3 =f(x 1,X 2 ) X 3 Func9on

16 Expression-based regulatory network inference Given A set of measured mrna levels across mul2ple biological samples Do Infer the regulators of genes Infer how regulators specify the expression of a gene Algorithms for network reconstruc2on vary based on their meaning of interac2on

17 Two classes of expression-based network inference methods Per-gene/direct methods X 1 X 2 X 1 X 4 X 3 X 5 Module based methods (Subsequent lectures) X 1 X 2 Module (Cluster) X 3 X 5

18 Per-gene methods X 1 X 2 X 3 X 4 X 3 X 5 Key idea: find the regulators that best explain expression level of a gene Probabilis2c graphical methods Bayesian network Sparse Candidates Dependency networks GENIE3, TIGRESS Informa2on theore2c methods Context Likelihood of relatedness ARACNE

19 Probabilis9c graphical models (PGMs) A marriage between probability and graph theory Nodes on the graph represent random variables Graph structure specifies sta2s2cal dependency structure Graph parameters specify the nature of the dependency PGMs can be directed or undirected Examples of PGMs: Bayesian networks, Dependency networks, Markov networks, Factor graphs

20 Different types of probabilis9c graphs In each graph type we can assert different condi2onal independencies Correla2on networks Gaussian Graphical models Dependency networks Bayesian networks

21 An undirected graph Correla9onal networks Edges represent high correla2on Need to determine what high is Edge weights denote different values of correla2on Cannot discriminate between direct and indirect correla2ons Random variables represent gene expression levels Msb2 Sho1 X 1 X 2 X 1 X 2 w 13 X 3 w 23 A measure of sta2s2cal correla2on (e.g. Pearson s correla2on) Ste20 X 3 An undirected weighted graph.

22 Popular examples of correla9onal networks Weighted Gene Co-expression Network Analysis (WGCNA) - Zhang and Horvath 2005 Relevance Networks BuBe & Kohane, 2000 Pacific symposium of biocompu2ng

23 Limita9ons of correla9onal networks Correla2onal networks cannot dis2nguish between direct and indirect dependencies This makes them less interpretable than other PGMs Coexpression Regulatory network Y Y Y H X Z X Z X Z X Y Z For any co-expression network, there are several possible regulatory networks that can explain these correla2ons. What we would like is to be able to discriminate between direct and indirect dependencies Here we need to review condi2onal independencies

24 Condi9onal independencies in PGMs The different classes of models we will see are based on a general no2on of specifying sta2s2cal independence Suppose we have two genes X and Y. We add an edge between X and Y if X and Y are not independent given a third set Z. Depending upon Z we will have a family of different PGMs

25 Condi9onal independence and PGMs Correla2onal networks Z is the empty set Markov networks X and Y are not independent given all other variables Gaussian Graphical models are a special case (later lectures) Dependency networks Approximate Markov networks May not be associated with a valid joint distribu2on (later lectures) First-order condi2onal independence models Explain the correla2on between two variables by a third variable Bayesian networks Generalize first-order condi2onal independence models

26 Bayesian networks (BN) A special type of probabilis2c graphical model Has two parts: A graph which is directed and acyclic A set of condi2onal distribu2ons Directed Acyclic Graph (DAG) The nodes denote random variables X 1 X N The edges encode sta2s2cal dependencies between the random variables establish parent child rela2onships Each node X i has a condi=onal probability distribu=on (CPD) represen2ng P(X i Pa(X i )); Pa: Parents Provides a tractable way to represent large joint distribu2ons

27 Key ques9ons in Bayesian networks What do the CPDs look like? What independence asser2ons can be made in Bayesian networks? How can we learn them from expression data? Inherent limita2ons of Bayesian networks

28 C An example Bayesian network P(S=F) P(S=T) F T Sprinkler (S) P(C=F) P(C=T) Cloudy (C) Rain (R) C P(R=F) P(R=T) F T WetGrass (W) S R P(W=F) P(W=T) F F 1 0 T F F T T T Adapted from Kevin Murphy: Intro to Graphical models and Bayes networks: hbp:// bnintro.html

29 Nota9on X i : i th random variable If there are few random variables, we will just use upper case lebers. E.g. A, B, C.. X={X 1,.., X p }: set of p random variables x ik : An assignment of X i in the kth sample x -ik : Set of assignments to all variables other than X i in the k th sample I(X i ;X j X k,x l ): Condi2onal independence nota2on: X i is independent of X j given X k and X l

30 Bayesian networks compactly represent joint distribu9ons P (X 1,,X p )= py P (X i Pa(X i )) i=1

31 Example Bayesian network of 5 variables PARENTS X 1 X 2 CHILD X 3 X 4 X 5 Assume X i is binary No independence asser2ons P (X) =P (X 1,X 2,X 3,X 4,X 5 ) Needs 2 5 measurements Independence asser2ons Needs 2 3 measurements P (X) =P (X 1 )P (X 2 )P (X 4 )P (X 3 X 1,X 2 )P (X 5 X 3,X 4 )

32 CPD in Bayesian networks The CPD P(X i Pa(X i )) specifies a distribu2on over values of X i for each combina2on of values of Pa(X i ) CPD P(X i Pa(X i )) can be parameterized in different ways X i are discrete random variables Condi2onal probability table or tree X i are con2nuous random variables CPD can be condi2onal Gaussians or regression trees

33 Represen9ng CPDs as tables Consider four binary variables X 1, X 2, X 3, X 4 P( X 4 X 1, X 2, X 3 ) as a table X 4 X 1 X 2 X 3 t f X 1 X 2 X 4 X 3 t t t t t f t f t t f f Pa(X 4 ): X 1, X 2, X 3 f t t f t f f f t f f f

34 Es9ma9ng CPD table from data Assume we observe the following assignments for X 1, X 2, X 3, X 4 N=7 T F T T For each joint assignment to X 1, X 2, X 3, X 1 X 2 X 3 X 4 T T F T es2mate the probabili2es for each value of X 4 T T F T T F T T T F T F T F T F F F T F For example, consider X 1 =T, X 2 =F, X 3 =T P(X 4 =T X 1 =T, X 2 =F, X 3 =T)=2/4 P(X 4 =F X 1 =T, X 2 =F, X 3 =T)=2/4

35 Condi9onal independencies in BN A variable X i is independent of its nondescendants given its parents I(X i ;X j X k,x l ): X i is independent of X j given X k and X l Consider the example Bayesian network. What are the set of condi2onal independencies in this graph? I(A; E) I(B; D A, E) I(D; E,B,C A) I(C; E, A, D B) I(E; A, D) Friedman et al., 2000

36 Bayesian network representa9on of a regulatory network Hot1 Sko1 HSP12 Inside the cell Hot1: Sko1: X 1 X 2 Random variables REGULATORS (PARENTS) P(X 1 ) X 1 X 2 P(X 2 ) Hsp12: X 3 TARGET (CHILD) X 3 P(X 3 X 1,X 2 ) Bayesian network

37 Learning problems in Bayesian networks Parameter learning on known graph structure Given a set of joint assignments of the random variables, es2mate the parameters of the model Structure learning Given a set of joint assignments of the random variables, es2mate the structure and parameters of the model Structure learning subsumes parameter learning

38 Network inference/graph structure learning is a computa9onally difficult problem Given 2 TFs and 3 nodes how many possible networks can there be? Not exhaus2ve set of possible networks. There can be a total of 2 6 possible networks.

39 Structure learning using score-based search Score(B) Describes how well B describes the data... Score(B 1 ) Score(B 2 ) Score(B 3 ) Score(B m ) Exhaus2ve search is not computa2onally tractable

40 Scoring a Bayesian network Maximum likelihood score Score ML (G) = max P (D G, ) Bayesian score Score Bayes (G) =P (D G) = P (D G)P (G) P (D) We typically ignore the denominator as it is the same for all models

41 Greedy Hill climbing to search Bayesian network space Input: Data D, An ini2al Bayesian network, B 0 ={G 0, Θ 0 } Output: B best Loop for r=1, 2.. un2l convergence: {B r1,.., B rm } = Neighbors(B r ) by making local changes to B r B r+1 : arg max j (Score(B rj )) Termina2on: B best = B r

42 Local changes to B i Current network D B C A B r add an edge D delete an edge D B C B C A B 1 Check for cycles A B 2 r r

43 Challenges with applying Bayesian network to genome-scale data Number of variables, p is in thousands Number of samples, N is in hundreds

44 Bayesian network-based methods to handle genome-scale networks Sparse candidate algorithm Friedman, Nachman, Pe er Friedman, Linial, Nachman, Pe er Module networks Segal, Pe er, Regev, Koller, Friedman. 2005

45 The Sparse candidate Structure learning in Bayesian networks A fast Bayesian network learning algorithm Key idea: Iden2fy k promising candidate parents for each X i k<<p, p: number of random variables Candidates define a skeleton graph H Restrict graph structure to select parents from H Early choices in H might exclude other good parents Resolve using an itera2ve algorithm

46 Sparse candidate algorithm Input: A data set D An ini2al Bayes net B 0 A parameter k: max number of parents per variable Output: Final B r Loop for r=1,2.. un2l convergence Restrict Based on D and B r-1 select candidate parents C i r for X i This defines a skeleton directed network H r Maximize Find network B r that maximizes the score Score(B r ) among networks sa2sfying Termina2on: Return B r Pa r (X i ) C r i

47 Informa9on theory for measuring dependence I(X;Y) is the mutual informa2on between two variables Knowing X, how much informa2on do we have for Y P(Z) is the probability distribu2on of Z Measures the difference between the two distribu2ons: joint and product of marginals

48 Selec9ng candidate parents in the Restrict Step A good parent for X i is one with strong sta2s2cal dependence with X i Mutual informa2on provides a good measure of sta2s2cal dependence I(X i ; X j ) Mutual informa2on should be used only as a first approxima2on Candidate parents need to be itera2vely refined to avoid missing important dependences A good parent for X i has the highest score improvement when added to Pa(X i )

49 Sparse candidate learns good networks faster than hill-climbing Greedy hill climbing takes much longer to reach a high scoring bayesian network Score (higher is beber) Greedy HC Disc Disc 15 Score 5 Score 15 Shld 5 Shld Time Time Greedy HC Disc 5 Disc 15 Score 5 Score 15 Shld 5 Shld 15 Dataset 1 Dataset variables 200 variables

50 Some comments about choosing candidates How to select k in the sparse candidate algorithm? Should k be the same for all X i? Regularized regression approaches can be used to es2mate the structure of an undirected graph Schmidt, Niculescu-Mizil, Murphy 2007 Es2mate an undirected dependency network Learn a Bayesian network constrained on the dependency network structure

51 Assessing confidence in the learned network Typically the number of training samples is not sufficient to reliably determine the right network One can however es2mate the confidence of specific features of the network Graph features f(g) Examples of f(g) An edge between two random variables Order rela2ons: Is X, Y s ancestor?

52 How to assess confidence in graph features? What we want is P(f(G) D), which is Gf(G)P (G D) But it is not feasible to compute this sum Instead we will use a bootstrap procedure

53 Bootstrap to assess graph feature confidence For i=1 to m Construct dataset D i by sampling with replacement N samples from dataset D, where N is the size of the original D Learn a network B i ={G i, Θ i } For each feature of interest f, calculate confidence Conf(f) = 1 mx f(g i ) m i=1

54 Does the bootstrap confidence represent real rela9onships? Compare the confidence distribu2on to that obtained from randomized data Shuffle the columns of each row (gene) separately Repeat the bootstrap procedure g Experimental condi2ons 1,1 g 1, n genes g2,1 g 2, n gm,1 g m, n randomize each row independently

55 Bootstrap-based confidence differs between real and actual data f Features with Confidence above Random Real f Friedman et al 2000

56 Example of a high confidence sub-network One learned Bayesian network Bootstrapped confidence Bayesian network: highlights a subnetwork associated with yeast ma2ng pathway. Colors indicate genes with known func2ons. Nir Friedman, Science 2004

57 Limita9ons with Bayesian networks Cannot model cyclic dependencies In prac2ce have not been shown to be beber than dependency networks However, most of the evalua2on has been done on structure not func2on Direc2onality is oxen not associated with causality Too many hidden variables in biological systems

58 A non-exhaus9ve list of expression-based network inference method Method Name Per-module Per-gene Parameters Model type Sparse candidate Bayesian network CLR Informa2on theore2c ARACNE Informa2on theore2c TIGRESS Dependency network Inferelator Dependency network GENIE3 Dependency network ModuleNetworks Bayesian network LemonTree Dependency network WGCNA Correla2on

59 ARACNE Es2mate mutual informa2on between all pairs of nodes Get rid of poten2al indirect links: These typically correspond to lowest mutual informa2on in a triple True network I(X 1,X 2 ) X 1 I(X 1,X 3 ) X 1 X 2 X 3 X 2 X 3 Regulators Target I(X 2,X 3 ) The edge between X 1 and X 3 would I(X 1,X 3 ) < min(i(x 1,X 2 ),I(X 2,X 3 )) have a low mutual informa2on Margolin et al 2006 Exclude edges with lowest informa2on in a triplet

60 Context Likelihood of Relatedness (CLR) M ij : Mutual informa2on between gene j and regulator i Context for gene j : mutual informa2on of j with all other regulators Context for regulator i : mutual informa2on of i with all other target genes Use the contexts to compute two background distribu2ons of mutual informa2on Get a z-value for M ij with respect to these distribu2ons. Combine regulator and target gene z-values into a single z-value Call an edge if z-value is greater than a cutoff. Faith et al, 2007

61 Using Context in CLR to prune out edges i M ij j z ij is the likelihood of observing M ij from either distribu2on by chance Use z ij to decide if gene i regulates gene j.

62 Plan for next lectures Dependency networks GENIE3 Evalua2on of different network inference algorithms Per-module networks Module networks Strengths and weaknesses of different methods Combining per-module and per-gene methods Case studies of expression-based network inference methods

63 References Kim, Harold D., Tal Shay, Erin K. O'Shea, and Aviv Regev. "Transcrip2onal Regulatory Circuits: Predic2ng Numbers from Alphabets." Science 325 (July 2009): De Smet, Riet and Kathleen Marchal. "Advantages and limita2ons of current network inference methods.." Nature reviews. Microbiology 8 (October 2010): Markowetz, Florian and Rainer Spang. "Inferring cellular networks-a review.." BMC bioinforma=cs 8 Suppl 6 (2007): S5+. N. Friedman, M. Linial, I. Nachman, and D. Pe'er, "Using bayesian networks to analyze expression data," Journal of Computa=onal Biology, vol. 7, no. 3-4, pp , Aug [Online]. Available: hbp:// dx.doi.org/ /

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models 02-710 Computational Genomics Systems biology Putting it together: Data integration using graphical models High throughput data So far in this class we discussed several different types of high throughput

More information

Priors in Dependency network learning

Priors in Dependency network learning Priors in Dependency network learning Sushmita Roy sroy@biostat.wisc.edu Computa:onal Network Biology Biosta2s2cs & Medical Informa2cs 826 Computer Sciences 838 hbps://compnetbiocourse.discovery.wisc.edu

More information

Learning in Bayesian Networks

Learning in Bayesian Networks Learning in Bayesian Networks Florian Markowetz Max-Planck-Institute for Molecular Genetics Computational Molecular Biology Berlin Berlin: 20.06.2002 1 Overview 1. Bayesian Networks Stochastic Networks

More information

Gene Regulatory Networks II Computa.onal Genomics Seyoung Kim

Gene Regulatory Networks II Computa.onal Genomics Seyoung Kim Gene Regulatory Networks II 02-710 Computa.onal Genomics Seyoung Kim Goal: Discover Structure and Func;on of Complex systems in the Cell Identify the different regulators and their target genes that are

More information

CSCI1950 Z Computa3onal Methods for Biology Lecture 24. Ben Raphael April 29, hgp://cs.brown.edu/courses/csci1950 z/ Network Mo3fs

CSCI1950 Z Computa3onal Methods for Biology Lecture 24. Ben Raphael April 29, hgp://cs.brown.edu/courses/csci1950 z/ Network Mo3fs CSCI1950 Z Computa3onal Methods for Biology Lecture 24 Ben Raphael April 29, 2009 hgp://cs.brown.edu/courses/csci1950 z/ Network Mo3fs Subnetworks with more occurrences than expected by chance. How to

More information

GLOBEX Bioinformatics (Summer 2015) Genetic networks and gene expression data

GLOBEX Bioinformatics (Summer 2015) Genetic networks and gene expression data GLOBEX Bioinformatics (Summer 2015) Genetic networks and gene expression data 1 Gene Networks Definition: A gene network is a set of molecular components, such as genes and proteins, and interactions between

More information

Computational Network Biology Biostatistics & Medical Informatics 826 Fall 2018

Computational Network Biology Biostatistics & Medical Informatics 826 Fall 2018 Computational Network Biology Biostatistics & Medical Informatics 826 Fall 2018 Sushmita Roy sroy@biostat.wisc.edu https://compnetbiocourse.discovery.wisc.edu Sep 6 th 2018 Goals for today Administrivia

More information

Bayesian networks Lecture 18. David Sontag New York University

Bayesian networks Lecture 18. David Sontag New York University Bayesian networks Lecture 18 David Sontag New York University Outline for today Modeling sequen&al data (e.g., =me series, speech processing) using hidden Markov models (HMMs) Bayesian networks Independence

More information

CSE 473: Ar+ficial Intelligence. Hidden Markov Models. Bayes Nets. Two random variable at each +me step Hidden state, X i Observa+on, E i

CSE 473: Ar+ficial Intelligence. Hidden Markov Models. Bayes Nets. Two random variable at each +me step Hidden state, X i Observa+on, E i CSE 473: Ar+ficial Intelligence Bayes Nets Daniel Weld [Most slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at hnp://ai.berkeley.edu.]

More information

Networks. Can (John) Bruce Keck Founda7on Biotechnology Lab Bioinforma7cs Resource

Networks. Can (John) Bruce Keck Founda7on Biotechnology Lab Bioinforma7cs Resource Networks Can (John) Bruce Keck Founda7on Biotechnology Lab Bioinforma7cs Resource Networks in biology Protein-Protein Interaction Network of Yeast Transcriptional regulatory network of E.coli Experimental

More information

Regulatory Inferece from Gene Expression. CMSC858P Spring 2012 Hector Corrada Bravo

Regulatory Inferece from Gene Expression. CMSC858P Spring 2012 Hector Corrada Bravo Regulatory Inferece from Gene Expression CMSC858P Spring 2012 Hector Corrada Bravo 2 Graphical Model Let y be a vector- valued random variable Suppose some condi8onal independence proper8es hold for some

More information

Gene Regula*on, ChIP- X and DNA Mo*fs. Statistics in Genomics Hongkai Ji

Gene Regula*on, ChIP- X and DNA Mo*fs. Statistics in Genomics Hongkai Ji Gene Regula*on, ChIP- X and DNA Mo*fs Statistics in Genomics Hongkai Ji (hji@jhsph.edu) Genetic information is stored in DNA TCAGTTGGAGCTGCTCCCCCACGGCCTCTCCTCACATTCCACGTCCTGTAGCTCTATGACCTCCACCTTTGAGTCCCTCCTC

More information

CS 6140: Machine Learning Spring What We Learned Last Week 2/26/16

CS 6140: Machine Learning Spring What We Learned Last Week 2/26/16 Logis@cs CS 6140: Machine Learning Spring 2016 Instructor: Lu Wang College of Computer and Informa@on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Sign

More information

Introduction to Bioinformatics

Introduction to Bioinformatics CSCI8980: Applied Machine Learning in Computational Biology Introduction to Bioinformatics Rui Kuang Department of Computer Science and Engineering University of Minnesota kuang@cs.umn.edu History of Bioinformatics

More information

CSCI1950 Z Computa3onal Methods for Biology* (*Working Title) Lecture 1. Ben Raphael January 21, Course Par3culars

CSCI1950 Z Computa3onal Methods for Biology* (*Working Title) Lecture 1. Ben Raphael January 21, Course Par3culars CSCI1950 Z Computa3onal Methods for Biology* (*Working Title) Lecture 1 Ben Raphael January 21, 2009 Course Par3culars Three major topics 1. Phylogeny: ~50% lectures 2. Func3onal Genomics: ~25% lectures

More information

CS839: Probabilistic Graphical Models. Lecture 2: Directed Graphical Models. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 2: Directed Graphical Models. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 2: Directed Graphical Models Theo Rekatsinas 1 Questions Questions? Waiting list Questions on other logistics 2 Section 1 1. Intro to Bayes Nets 3 Section

More information

Learning structurally consistent undirected probabilistic graphical models

Learning structurally consistent undirected probabilistic graphical models Learning structurally consistent undirected probabilistic graphical models Sushmita Roy Department of Computer Science University of New Mexico Terran Lane Department of Computer Science University of

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Inferring Transcriptional Regulatory Networks from High-throughput Data

Inferring Transcriptional Regulatory Networks from High-throughput Data Inferring Transcriptional Regulatory Networks from High-throughput Data Lectures 9 Oct 26, 2011 CSE 527 Computational Biology, Fall 2011 Instructor: Su-In Lee TA: Christopher Miles Monday & Wednesday 12:00-1:20

More information

Directed Graphical Models or Bayesian Networks

Directed Graphical Models or Bayesian Networks Directed Graphical Models or Bayesian Networks Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Bayesian Networks One of the most exciting recent advancements in statistical AI Compact

More information

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, Networks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 4 Learning Bayesian Networks CS/CNS/EE 155 Andreas Krause Announcements Another TA: Hongchao Zhou Please fill out the questionnaire about recitations Homework 1 out.

More information

Inferring Protein-Signaling Networks II

Inferring Protein-Signaling Networks II Inferring Protein-Signaling Networks II Lectures 15 Nov 16, 2011 CSE 527 Computational Biology, Fall 2011 Instructor: Su-In Lee TA: Christopher Miles Monday & Wednesday 12:00-1:20 Johnson Hall (JHN) 022

More information

Dynamic Approaches: The Hidden Markov Model

Dynamic Approaches: The Hidden Markov Model Dynamic Approaches: The Hidden Markov Model Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Inference as Message

More information

Computer Vision. Pa0ern Recogni4on Concepts Part I. Luis F. Teixeira MAP- i 2012/13

Computer Vision. Pa0ern Recogni4on Concepts Part I. Luis F. Teixeira MAP- i 2012/13 Computer Vision Pa0ern Recogni4on Concepts Part I Luis F. Teixeira MAP- i 2012/13 What is it? Pa0ern Recogni4on Many defini4ons in the literature The assignment of a physical object or event to one of

More information

Latent Variable models for GWAs

Latent Variable models for GWAs Latent Variable models for GWAs Oliver Stegle Machine Learning and Computational Biology Research Group Max-Planck-Institutes Tübingen, Germany September 2011 O. Stegle Latent variable models for GWAs

More information

3 : Representation of Undirected GM

3 : Representation of Undirected GM 10-708: Probabilistic Graphical Models 10-708, Spring 2016 3 : Representation of Undirected GM Lecturer: Eric P. Xing Scribes: Longqi Cai, Man-Chia Chang 1 MRF vs BN There are two types of graphical models:

More information

2 : Directed GMs: Bayesian Networks

2 : Directed GMs: Bayesian Networks 10-708: Probabilistic Graphical Models 10-708, Spring 2017 2 : Directed GMs: Bayesian Networks Lecturer: Eric P. Xing Scribes: Jayanth Koushik, Hiroaki Hayashi, Christian Perez Topic: Directed GMs 1 Types

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

CS 6140: Machine Learning Spring What We Learned Last Week. Survey 2/26/16. VS. Model

CS 6140: Machine Learning Spring What We Learned Last Week. Survey 2/26/16. VS. Model Logis@cs CS 6140: Machine Learning Spring 2016 Instructor: Lu Wang College of Computer and Informa@on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Assignment

More information

The Monte Carlo Method: Bayesian Networks

The Monte Carlo Method: Bayesian Networks The Method: Bayesian Networks Dieter W. Heermann Methods 2009 Dieter W. Heermann ( Methods)The Method: Bayesian Networks 2009 1 / 18 Outline 1 Bayesian Networks 2 Gene Expression Data 3 Bayesian Networks

More information

CS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 7: Learning Fully Observed BNs Theo Rekatsinas 1 Exponential family: a basic building block For a numeric random variable X p(x ) =h(x)exp T T (x) A( ) = 1

More information

CS 6140: Machine Learning Spring 2016

CS 6140: Machine Learning Spring 2016 CS 6140: Machine Learning Spring 2016 Instructor: Lu Wang College of Computer and Informa?on Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Email: luwang@ccs.neu.edu Logis?cs Assignment

More information

Bayesian Networks. Motivation

Bayesian Networks. Motivation Bayesian Networks Computer Sciences 760 Spring 2014 http://pages.cs.wisc.edu/~dpage/cs760/ Motivation Assume we have five Boolean variables,,,, The joint probability is,,,, How many state configurations

More information

Computational Genomics. Reconstructing dynamic regulatory networks in multiple species

Computational Genomics. Reconstructing dynamic regulatory networks in multiple species 02-710 Computational Genomics Reconstructing dynamic regulatory networks in multiple species Methods for reconstructing networks in cells CRH1 SLT2 SLR3 YPS3 YPS1 Amit et al Science 2009 Pe er et al Recomb

More information

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016

CS 2750: Machine Learning. Bayesian Networks. Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 CS 2750: Machine Learning Bayesian Networks Prof. Adriana Kovashka University of Pittsburgh March 14, 2016 Plan for today and next week Today and next time: Bayesian networks (Bishop Sec. 8.1) Conditional

More information

Representation of undirected GM. Kayhan Batmanghelich

Representation of undirected GM. Kayhan Batmanghelich Representation of undirected GM Kayhan Batmanghelich Review Review: Directed Graphical Model Represent distribution of the form ny p(x 1,,X n = p(x i (X i i=1 Factorizes in terms of local conditional probabilities

More information

Discovering molecular pathways from protein interaction and ge

Discovering molecular pathways from protein interaction and ge Discovering molecular pathways from protein interaction and gene expression data 9-4-2008 Aim To have a mechanism for inferring pathways from gene expression and protein interaction data. Motivation Why

More information

Graphical Models. Lecture 3: Local Condi6onal Probability Distribu6ons. Andrew McCallum

Graphical Models. Lecture 3: Local Condi6onal Probability Distribu6ons. Andrew McCallum Graphical Models Lecture 3: Local Condi6onal Probability Distribu6ons Andrew McCallum mccallum@cs.umass.edu Thanks to Noah Smith and Carlos Guestrin for some slide materials. 1 Condi6onal Probability Distribu6ons

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models David Sontag New York University Lecture 4, February 16, 2012 David Sontag (NYU) Graphical Models Lecture 4, February 16, 2012 1 / 27 Undirected graphical models Reminder

More information

Soft Computing. Lecture Notes on Machine Learning. Matteo Matteucci.

Soft Computing. Lecture Notes on Machine Learning. Matteo Matteucci. Soft Computing Lecture Notes on Machine Learning Matteo Matteucci matteucci@elet.polimi.it Department of Electronics and Information Politecnico di Milano Matteo Matteucci c Lecture Notes on Machine Learning

More information

Introduction to Probabilistic Graphical Models

Introduction to Probabilistic Graphical Models Introduction to Probabilistic Graphical Models Franz Pernkopf, Robert Peharz, Sebastian Tschiatschek Graz University of Technology, Laboratory of Signal Processing and Speech Communication Inffeldgasse

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Probability and Structure in Natural Language Processing

Probability and Structure in Natural Language Processing Probability and Structure in Natural Language Processing Noah Smith, Carnegie Mellon University 2012 Interna@onal Summer School in Language and Speech Technologies Quick Recap Yesterday: Bayesian networks

More information

STAD68: Machine Learning

STAD68: Machine Learning STAD68: Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! h0p://www.cs.toronto.edu/~rsalakhu/ Lecture 1 Evalua;on 3 Assignments worth 40%. Midterm worth 20%. Final

More information

A Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004

A Brief Introduction to Graphical Models. Presenter: Yijuan Lu November 12,2004 A Brief Introduction to Graphical Models Presenter: Yijuan Lu November 12,2004 References Introduction to Graphical Models, Kevin Murphy, Technical Report, May 2001 Learning in Graphical Models, Michael

More information

Directed and Undirected Graphical Models

Directed and Undirected Graphical Models Directed and Undirected Graphical Models Adrian Weller MLSALT4 Lecture Feb 26, 2016 With thanks to David Sontag (NYU) and Tony Jebara (Columbia) for use of many slides and illustrations For more information,

More information

Transcrip:on factor binding mo:fs

Transcrip:on factor binding mo:fs Transcrip:on factor binding mo:fs BMMB- 597D Lecture 29 Shaun Mahony Transcrip.on factor binding sites Short: Typically between 6 20bp long Degenerate: TFs have favorite binding sequences but don t require

More information

CSE 473: Ar+ficial Intelligence. Probability Recap. Markov Models - II. Condi+onal probability. Product rule. Chain rule.

CSE 473: Ar+ficial Intelligence. Probability Recap. Markov Models - II. Condi+onal probability. Product rule. Chain rule. CSE 473: Ar+ficial Intelligence Markov Models - II Daniel S. Weld - - - University of Washington [Most slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188

More information

CSE 473: Ar+ficial Intelligence

CSE 473: Ar+ficial Intelligence CSE 473: Ar+ficial Intelligence Hidden Markov Models Luke Ze@lemoyer - University of Washington [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188

More information

COMP538: Introduction to Bayesian Networks

COMP538: Introduction to Bayesian Networks COMP538: Introduction to Bayesian Networks Lecture 9: Optimal Structure Learning Nevin L. Zhang lzhang@cse.ust.hk Department of Computer Science and Engineering Hong Kong University of Science and Technology

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Undirected Graphical Models Mark Schmidt University of British Columbia Winter 2016 Admin Assignment 3: 2 late days to hand it in today, Thursday is final day. Assignment 4:

More information

Bayesian Networks Inference with Probabilistic Graphical Models

Bayesian Networks Inference with Probabilistic Graphical Models 4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning

More information

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, Networks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Outline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination

Outline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination Probabilistic Graphical Models COMP 790-90 Seminar Spring 2011 The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Outline It Introduction ti Representation Bayesian network Conditional Independence Inference:

More information

Inferring Transcriptional Regulatory Networks from Gene Expression Data II

Inferring Transcriptional Regulatory Networks from Gene Expression Data II Inferring Transcriptional Regulatory Networks from Gene Expression Data II Lectures 9 Oct 26, 2011 CSE 527 Computational Biology, Fall 2011 Instructor: Su-In Lee TA: Christopher Miles Monday & Wednesday

More information

Lecture 6: Graphical Models

Lecture 6: Graphical Models Lecture 6: Graphical Models Kai-Wei Chang CS @ Uniersity of Virginia kw@kwchang.net Some slides are adapted from Viek Skirmar s course on Structured Prediction 1 So far We discussed sequence labeling tasks:

More information

Boolean Gossip Networks

Boolean Gossip Networks Boolean Gossip Networks Guodong Shi Research School of Engineering The Australian Na

More information

10708 Graphical Models: Homework 2

10708 Graphical Models: Homework 2 10708 Graphical Models: Homework 2 Due Monday, March 18, beginning of class Feburary 27, 2013 Instructions: There are five questions (one for extra credit) on this assignment. There is a problem involves

More information

Learning structurally consistent undirected probabilistic graphical models

Learning structurally consistent undirected probabilistic graphical models Learning structurally consistent undirected probabilistic graphical models Sushmita Roy sroy@cs.unm.edu Terran Lane terran@cs.unm.edu Margaret Werner-Washburne + maggieww@unm.edu Dept. of Computer Science

More information

Review: Bayesian learning and inference

Review: Bayesian learning and inference Review: Bayesian learning and inference Suppose the agent has to make decisions about the value of an unobserved query variable X based on the values of an observed evidence variable E Inference problem:

More information

Computational Genomics

Computational Genomics Computational Genomics http://www.cs.cmu.edu/~02710 Introduction to probability, statistics and algorithms (brief) intro to probability Basic notations Random variable - referring to an element / event

More information

Mathematical Formulation of Our Example

Mathematical Formulation of Our Example Mathematical Formulation of Our Example We define two binary random variables: open and, where is light on or light off. Our question is: What is? Computer Vision 1 Combining Evidence Suppose our robot

More information

Learning Causality. Sargur N. Srihari. University at Buffalo, The State University of New York USA

Learning Causality. Sargur N. Srihari. University at Buffalo, The State University of New York USA Learning Causality Sargur N. Srihari University at Buffalo, The State University of New York USA 1 Plan of Discussion Bayesian Networks Causal Models Learning Causal Models 2 BN and Complexity of Prob

More information

Sta$s$cal sequence recogni$on

Sta$s$cal sequence recogni$on Sta$s$cal sequence recogni$on Determinis$c sequence recogni$on Last $me, temporal integra$on of local distances via DP Integrates local matches over $me Normalizes $me varia$ons For cts speech, segments

More information

Bioinformatics 2 - Lecture 4

Bioinformatics 2 - Lecture 4 Bioinformatics 2 - Lecture 4 Guido Sanguinetti School of Informatics University of Edinburgh February 14, 2011 Sequences Many data types are ordered, i.e. you can naturally say what is before and what

More information

CSCI1950 Z Computa4onal Methods for Biology Lecture 4. Ben Raphael February 2, hhp://cs.brown.edu/courses/csci1950 z/ Algorithm Summary

CSCI1950 Z Computa4onal Methods for Biology Lecture 4. Ben Raphael February 2, hhp://cs.brown.edu/courses/csci1950 z/ Algorithm Summary CSCI1950 Z Computa4onal Methods for Biology Lecture 4 Ben Raphael February 2, 2009 hhp://cs.brown.edu/courses/csci1950 z/ Algorithm Summary Parsimony Probabilis4c Method Input Output Sankoff s & Fitch

More information

Lecture 4 October 18th

Lecture 4 October 18th Directed and undirected graphical models Fall 2017 Lecture 4 October 18th Lecturer: Guillaume Obozinski Scribe: In this lecture, we will assume that all random variables are discrete, to keep notations

More information

Temporal Models.

Temporal Models. Temporal Models www.cs.wisc.edu/~page/cs760/ Goals for the lecture (with thanks to Irene Ong, Stuart Russell and Peter Norvig, Jakob Rasmussen, Jeremy Weiss, Yujia Bao, Charles Kuang, Peggy Peissig, and

More information

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network. ecall from last time Lecture 3: onditional independence and graph structure onditional independencies implied by a belief network Independence maps (I-maps) Factorization theorem The Bayes ball algorithm

More information

Probabilistic graphical models CS 535c (Topics in AI) Stat 521a (Topics in multivariate analysis) Lecture 1. Kevin Murphy

Probabilistic graphical models CS 535c (Topics in AI) Stat 521a (Topics in multivariate analysis) Lecture 1. Kevin Murphy Probabilistic graphical models CS 535c (Topics in AI) Stat 521a (Topics in multivariate analysis) Lecture 1 Kevin Murphy Monday 13 September, 2004 Administrivia Lectures: MW 9.30-10.50, CISR 304 Regular

More information

CSCI 360 Introduc/on to Ar/ficial Intelligence Week 2: Problem Solving and Op/miza/on

CSCI 360 Introduc/on to Ar/ficial Intelligence Week 2: Problem Solving and Op/miza/on CSCI 360 Introduc/on to Ar/ficial Intelligence Week 2: Problem Solving and Op/miza/on Professor Wei-Min Shen Week 13.1 and 13.2 1 Status Check Extra credits? Announcement Evalua/on process will start soon

More information

Introduction to Probabilistic Graphical Models

Introduction to Probabilistic Graphical Models Introduction to Probabilistic Graphical Models Kyu-Baek Hwang and Byoung-Tak Zhang Biointelligence Lab School of Computer Science and Engineering Seoul National University Seoul 151-742 Korea E-mail: kbhwang@bi.snu.ac.kr

More information

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering Types of learning Modeling data Supervised: we know input and targets Goal is to learn a model that, given input data, accurately predicts target data Unsupervised: we know the input only and want to make

More information

Inferring Protein-Signaling Networks

Inferring Protein-Signaling Networks Inferring Protein-Signaling Networks Lectures 14 Nov 14, 2011 CSE 527 Computational Biology, Fall 2011 Instructor: Su-In Lee TA: Christopher Miles Monday & Wednesday 12:00-1:20 Johnson Hall (JHN) 022 1

More information

Phylogene)cs. IMBB 2016 BecA- ILRI Hub, Nairobi May 9 20, Joyce Nzioki

Phylogene)cs. IMBB 2016 BecA- ILRI Hub, Nairobi May 9 20, Joyce Nzioki Phylogene)cs IMBB 2016 BecA- ILRI Hub, Nairobi May 9 20, 2016 Joyce Nzioki Phylogenetics The study of evolutionary relatedness of organisms. Derived from two Greek words:» Phle/Phylon: Tribe/Race» Genetikos:

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Matrix Data: Classification: Part 2 Instructor: Yizhou Sun yzsun@ccs.neu.edu September 21, 2014 Methods to Learn Matrix Data Set Data Sequence Data Time Series Graph & Network

More information

Probabilistic Graphical Models (I)

Probabilistic Graphical Models (I) Probabilistic Graphical Models (I) Hongxin Zhang zhx@cad.zju.edu.cn State Key Lab of CAD&CG, ZJU 2015-03-31 Probabilistic Graphical Models Modeling many real-world problems => a large number of random

More information

Chris Bishop s PRML Ch. 8: Graphical Models

Chris Bishop s PRML Ch. 8: Graphical Models Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular

More information

Latent Dirichlet Alloca/on

Latent Dirichlet Alloca/on Latent Dirichlet Alloca/on Blei, Ng and Jordan ( 2002 ) Presented by Deepak Santhanam What is Latent Dirichlet Alloca/on? Genera/ve Model for collec/ons of discrete data Data generated by parameters which

More information

2 : Directed GMs: Bayesian Networks

2 : Directed GMs: Bayesian Networks 10-708: Probabilistic Graphical Models, Spring 2015 2 : Directed GMs: Bayesian Networks Lecturer: Eric P. Xing Scribes: Yi Cheng, Cong Lu 1 Notation Here the notations used in this course are defined:

More information

Bayesian Networks to design optimal experiments. Davide De March

Bayesian Networks to design optimal experiments. Davide De March Bayesian Networks to design optimal experiments Davide De March davidedemarch@gmail.com 1 Outline evolutionary experimental design in high-dimensional space and costly experimentation the microwell mixture

More information

Sum-Product Networks: A New Deep Architecture

Sum-Product Networks: A New Deep Architecture Sum-Product Networks: A New Deep Architecture Pedro Domingos Dept. Computer Science & Eng. University of Washington Joint work with Hoifung Poon 1 Graphical Models: Challenges Bayesian Network Markov Network

More information

Written Exam 15 December Course name: Introduction to Systems Biology Course no

Written Exam 15 December Course name: Introduction to Systems Biology Course no Technical University of Denmark Written Exam 15 December 2008 Course name: Introduction to Systems Biology Course no. 27041 Aids allowed: Open book exam Provide your answers and calculations on separate

More information

Introduction to Bioinformatics. Shifra Ben-Dor Irit Orr

Introduction to Bioinformatics. Shifra Ben-Dor Irit Orr Introduction to Bioinformatics Shifra Ben-Dor Irit Orr Lecture Outline: Technical Course Items Introduction to Bioinformatics Introduction to Databases This week and next week What is bioinformatics? A

More information

Abstract. Three Methods and Their Limitations. N-1 Experiments Suffice to Determine the Causal Relations Among N Variables

Abstract. Three Methods and Their Limitations. N-1 Experiments Suffice to Determine the Causal Relations Among N Variables N-1 Experiments Suffice to Determine the Causal Relations Among N Variables Frederick Eberhardt Clark Glymour 1 Richard Scheines Carnegie Mellon University Abstract By combining experimental interventions

More information

Directed and Undirected Graphical Models

Directed and Undirected Graphical Models Directed and Undirected Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Last Lecture Refresher Lecture Plan Directed

More information

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4 ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian

More information

Lecture 5: Bayesian Network

Lecture 5: Bayesian Network Lecture 5: Bayesian Network Topics of this lecture What is a Bayesian network? A simple example Formal definition of BN A slightly difficult example Learning of BN An example of learning Important topics

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information

Predicting causal effects in large-scale systems from observational data

Predicting causal effects in large-scale systems from observational data nature methods Predicting causal effects in large-scale systems from observational data Marloes H Maathuis 1, Diego Colombo 1, Markus Kalisch 1 & Peter Bühlmann 1,2 Supplementary figures and text: Supplementary

More information

Lecture 6: Graphical Models: Learning

Lecture 6: Graphical Models: Learning Lecture 6: Graphical Models: Learning 4F13: Machine Learning Zoubin Ghahramani and Carl Edward Rasmussen Department of Engineering, University of Cambridge February 3rd, 2010 Ghahramani & Rasmussen (CUED)

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 11 CRFs, Exponential Family CS/CNS/EE 155 Andreas Krause Announcements Homework 2 due today Project milestones due next Monday (Nov 9) About half the work should

More information

Inference in Graphical Models Variable Elimination and Message Passing Algorithm

Inference in Graphical Models Variable Elimination and Message Passing Algorithm Inference in Graphical Models Variable Elimination and Message Passing lgorithm Le Song Machine Learning II: dvanced Topics SE 8803ML, Spring 2012 onditional Independence ssumptions Local Markov ssumption

More information

Part 2. Representation Learning Algorithms

Part 2. Representation Learning Algorithms 53 Part 2 Representation Learning Algorithms 54 A neural network = running several logistic regressions at the same time If we feed a vector of inputs through a bunch of logis;c regression func;ons, then

More information

Directed Graphical Models

Directed Graphical Models Directed Graphical Models Instructor: Alan Ritter Many Slides from Tom Mitchell Graphical Models Key Idea: Conditional independence assumptions useful but Naïve Bayes is extreme! Graphical models express

More information

Physical network models and multi-source data integration

Physical network models and multi-source data integration Physical network models and multi-source data integration Chen-Hsiang Yeang MIT AI Lab Cambridge, MA 02139 chyeang@ai.mit.edu Tommi Jaakkola MIT AI Lab Cambridge, MA 02139 tommi@ai.mit.edu September 30,

More information

Inferring Causal Phenotype Networks from Segregating Populat

Inferring Causal Phenotype Networks from Segregating Populat Inferring Causal Phenotype Networks from Segregating Populations Elias Chaibub Neto chaibub@stat.wisc.edu Statistics Department, University of Wisconsin - Madison July 15, 2008 Overview Introduction Description

More information

25 : Graphical induced structured input/output models

25 : Graphical induced structured input/output models 10-708: Probabilistic Graphical Models 10-708, Spring 2016 25 : Graphical induced structured input/output models Lecturer: Eric P. Xing Scribes: Raied Aljadaany, Shi Zong, Chenchen Zhu Disclaimer: A large

More information