Dobbiaco Lectures 2010 (2)Solved and Unsolved Problems in

Size: px
Start display at page:

Download "Dobbiaco Lectures 2010 (2)Solved and Unsolved Problems in"

Transcription

1 Dobbiaco Lectures 2010 (2) Solved and Unsolved Problems in Biology Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Dobbiaco

2 Outline

3 PART III : Uncertainty

4 Main theses Outline A thought is a proposition with sense. 3 A proposition is a truth-function of elementary propositions. 4 The general form of a proposition is the general form of a truth function, which is: p, ξ, ξ 5 Where (or of what) one cannot speak, one must pass over in silence. Ludwig Wittgenstein, Tractatus Logico-Philosophicus, 1921.

5 Outline

6 Random Variables A (discrete) random variable is a numerical quantity that in some experiment (involving randomness) takes a value from some (discrete) set of possible values. More formally, these are measurable maps X(ω),ω Ω, from a basic probability space (Ω, F, P) ( outcomes, a sigma field of subsets of Ω and probability measure P on F ). Events...{ω Ω X(ω) = x i }... same as {X = x i } [X assumes the value x i ].

7 Few Examples Example 1: Rolling of two six-sided dice. Random Variable might be the sum of the two numbers showing on the dice. The possible values of the random variable are 2, 3,..., 12. Example 2: Occurrence of a specific word GAATTC in a genome. Random Variable might be the number of occurrence of this word in a random genome of length The possible values of the random variable are 0, 1, 2,...,

8 The Probability Distribution The probability distribution of a discrete random variable Y is the set of values that this random variable can take, together with the set of associated probabilities. Probabilities are numbers in the range between zero and one (inclusive) that always add up to one when summed over all possible values of the random variable.

9 Bernoulli Trial Outline A Bernoulli trial is a single trial with two possible outcomes: success & failure. P(success) = p and P(failure) = 1 p q. Random variable S takes the value 1 if the trial results in failure and +1 if it results in success. P S (s) = p (1+s)/2 q (1 s)/2, s = 1,+1.

10 The Binomial Distribution A Binomial random variable is the number of successes in a fixed number n of independent Bernoulli trials (with success probability = p). Random variable Y denotes the total number of successes in the n trials. ( ) n P Y (y) = p y q n y, y = 0, 1,...,n. y

11 The Poisson Distribution A random variable Y has a Poisson distribution (with parameter λ > 0) if P Y (y) = e λ λ y, y = 0, 1,... y! The Poisson distribution often arises as a limiting form of the binomial distribution.

12 Continuous Random Variables We denote a continuous random variable by X and observed value of the random variable by x. Each random variable X with range I has an associated density function f X (x) which is defined, positive for all x and integrates to one over the range I. Prob(a < X < b) = b a f X (x)dx.

13 The Normal Distribution A random variable X has a normal or Gaussian distribution if it has range (, ) and density function f X (x) = 1 2πσ e (x µ)2 2σ 2, where µ and σ > 0 are parameters of the distribution.

14 Expectation Outline For a random variable Y, and any function g(y) of Y, the expected value of g(y) is E(g(Y)) = y g(y)p Y (y), when Y is discrete; and E(g(Y)) = y g(y)f Y (y) dy, when Y is continuous. Thus, mean(y) = E(Y) = µ(y), variance(y) = E(Y 2 ) E(Y) 2 = σ 2 (Y).

15 Conditional Probabilities Suppose that A 1 and A 2 are two events such that P(A 2 ) 0. Then the conditional probability that the event A 1 occurs, given that event A 2 occurs, denoted by P(A 1 A 2 ) is given by the formula P(A 1 A 2 ) = P(A 1 A 2 ). P(A 2 )

16 Bayes Rule Outline Suppose that A 1 and A 2 are two events such that P(A 1 ) 0 and P(A 2 ) 0. Then P(A 2 A 1 ) = P(A 2)P(A 1 A 2 ). P(A 1 )

17 Bayes Nets Outline Bayes Nets or Bayesian networks are graphical representation for probabilistic relationships among a set of random variables. Given a finite set X = {X 1,..., X n } of discrete random variables where each variable X i may take values from a finite set, denoted by Val(X i ). A Bayesian network is an annotated directed acyclic graph (DAG) G that encodes a joint probability distribution over X.

18 The graph G (Bayesian Network) is defined as follows: The nodes of the graph correspond to the random variables X 1, X 2,...,X n. The links of the graph correspond to the direct influence from one variable to the other. 1 If there is a directed link from variablex i to variable X j, variable X i will be a parent of variable X j. 2 Each node is annotated with a conditional probability distribution (CPD) that represents P(X i Pa(X i )), where Pa(X i ) denotes the parents of X i in G.

19 The pair (G, CPD) encodes the joint distribution P(X 1,..., X n ). A unique joint probability distribution over X from G is factorized as: P(X 1,..., X n ) = i (P(X i Pa(X i ))).

20

21 P(X 1 = no) = 0.8 P(X 1 = yes) = 0.2 P(X 2 = absent X 1 = no) = 0.95 P(X 2 = absent X 1 = yes) = 0.75 P(X 2 = present X 1 = no) = 0.05 P(X 2 = present X 1 = yes) = 0.25 P(X 3 = absent X 1 = no) = P(X 3 = absent X 1 = yes) = P(X 3 = absent X 1 = no) = P(X 3 = absent X 1 = yes) = P(X 4 = absent X 2 = absent, P(X 4 = absent X 2 = absent, X 3 = absent) = 0.95 X 3 = present) = 0.5 P(X 4 = absent X 2 = present, P(X 4 = absent X 2 = present, X 3 = absent) = 0.9 X 3 = present) = 0.25 P(X 4 = present X 2 = absent, P(X 4 = present X 2 = absent, X 3 = absent) = 0.05 X 3 = present) = 0.5 P(X 4 = present X 2 = present, P(X 4 = present X 2 = present, X 3 = absent) = 0.1 X 3 = present) = 0.75 P(X 5 = absent X 3 = absent) = 0.98 P(X 5 = absent X 3 = present) = 0.4 P(X 5 = present X 3 = absent) = 0.02 P(X 5 = present X 3 = present) = 0.6

22 Causal Bayesian networks A causal Bayesian network of a domain is similar as the normal Bayesian network the difference is in the explanation of the links in the Bayesian networks. In the normal Bayesian networks, the links between variables can be explained as correlation or association. In a causal Bayesian network, the links mean that the parent variables will causally influence the values of the child variables.

23 The causal influence in this thesis is defined based on manipulation criteria: Manipulation Criteria Suppose there are two variables A and B in the domain; If we can manipulate the variables in the domain, set the value of variable A as a 1 or a 2 and measure its effect on variable B, then the probability distribution of variable B will change under the conditions of the different values of variable A P(B do(a = a 1 )) P(B do(a = a 2 )).

24 Bayesian network structure learning The main task in Bayesian network structure learning: Find a structure of Bayesian network that describes the observed data the best. The problem is NP-complete. Many heuristics have been proposed to learn Bayesian network structure. There are two categories of approaches for Bayesian network structure learning: The score-and-search-based approach and The constraint-based approach.

25 The score-and-search-based approach The score-and-search-based approach: The methods in this category start from an initial structure (generated randomly or from domain knowledge) and move to the neighbors with the best score in the structure space determinately or stochastically until a local maximum of the selected criteria is reached. The greedy learning process can re-start several times with different initial structures to improve the result.

26 The constraint-based approach The constraint-based approach: The methods under this category start to test the statistical significance of the pairs of variables conditioning on other variables to induce conditional independence. The pairs of variables which pass some threshold are deemed as directly connected in the Bayesian networks. The complete Bayesian network structure is constructed from the induced conditional independence and dependence information.

27 How can Bayes Nets be Causal? The causal Markov condition is a relative of Reichenbach s thesis that conditioning on common causes will render joint effects independent of other another. One can then add the assumption of faithfulness or stability as well as to assume that all underlying systems of causal laws are deterministic. Similarly, using causal minimality (or sufficiency) assumption, one may try to justify the claim that Bayesian Nets Are All There Is to Causality... Bayes nets encode information about probabilistic independencies. Causality, if it has any connection with probability, would seem to be related to probabilistic dependence. (Cartwright, N.)

28 Probability Raising Causes produce their effects; they make them happen. So, in the right kind of population we can expect that there will be a higher frequency of the effect (E) when the cause (C) is present than when it is absent; and conversely for preventatives. What kind of populations are the right kinds? Populations in which the requisite causal process sometimes operates unimpeded and its doing so is not correlated with other processes that mask the increase in probability, such as the presence of a process preventing the effect or the absence of another positive process. Cartwright

29 PART IV:

30 Outline

31

32 The law of causality... is a relic of a bygone age, surviving, like the monarchy, only because it is erroneously supposed to do no harm... Bertrand Russell, On the Notion of Cause. Proceedings of the Aristotelian Society 13: 1-26, 1913.

33 and Correlation A fallacy, known as cum hoc ergo propter hoc (Latin for with this, therefore because of this ): Correlations do not imply causation. The Probability Raising condition + Temporal Priority Condition

34 Regularity Theories (David Hume) Causes are invariably followed by their effects: We may define a cause to be an object, followed by another, and where all the objects similar to the first, are followed by objects similar to the second. Attempts to analyze causation in terms of invariable patterns of succession are referred to as regularity theories of causation. There are a number of well-known difficulties with regularity theories, and these may be used to motivate probabilistic approaches to causation.

35 Imperfect Regularities The first difficulty is that most causes are not invariably followed by their effects. Penetrance: The presence of a disease allele does not always lead to a disease phenotype. Probabilistic theories of causation: simply requires that causes raise the probability of their effects; an effect may still occur in the absence of a cause or fail to occur in its presence. Thus smoking is a cause of lung cancer, not because all smokers develop lung cancer, but because smokers are more likely to develop lung cancer than non-smokers.

36 Imperfect Regularities: INUS condition John Stuart Mill and John Mackie offer more refined accounts of the regularities that underwrite causal relations. An INUS condition: for some effect is an insufficient but non-redundant part of an unnecessary but sufficient condition. Complexity: raises problems for the epistemology of causation.

37 INUS condition Suppose, for example, that a lit match causes a forest fire. The lighting of the match, by itself, is not sufficient; many matches are lit without ensuing forest fires. The lit match is, however, a part of some constellation of conditions that are jointly sufficient for the fire. Moreover, given that this set of conditions occurred, rather than some other set sufficient for fire, the lighting of the match was necessary: fires do not occur in such circumstances when lit matches are not present. Epistasis, and gene-environment interaction.

38 Asymmetry Outline is usually asymmetric. If A causes B, then, typically, B will not also cause A. This poses a problem for regularity theories, for it seems quite plausible that if smoking is an INUS condition for lung cancer, then lung cancer will be an INUS condition for smoking. One way of enforcing the asymmetry of causation is to stipulate that causes precede their effects in time.

39 Spurious Regularities Suppose that a cause is regularly followed by two effects. For instance, a particular allele A is pleiotropic... It causes a disease trait, but also transcription of another gene B. B may be mistakenly thought to be causing the disease. B is also an INUS condition for disease state. But it s not a cause. Whenever the barometric pressure drops below a certain level, two things happen: First, the height of the column of mercury in a barometer drops. Shortly afterwards, a storm occurs. Then, it may well also be the case that whenever the column of mercury drops, there will be a storm.

40 Causes raise the probability of their effects. This can be expressed formally using the apparatus of conditional probability. Let A, B, C,... represent factors that potentially stand in causal relations. Let P be a probability function... such that Pr(A) represents the empirical probability that factor A occurs or is instantiated. Recall: P(B A) represents the conditional probability of B, given A. Pr(A B) Pr(B A) =. Pr(A)

41 If Pr(A) is 0, then the ratio in the definition of conditional probability is undefined. (There are other ways of handling this formally). A raises the probability of B is that Pr(B A) > Pr(B A). PR Axiom PR: A causes B if and only ifpr(b A) > Pr(B A).

42 Problems Outline Probability-raising is symmetric: if Pr(B A) > P(B A), then Pr(A B) > P(A B). The causal relation, however, is typically asymmetric. Probability-raising has trouble with spurious correlations. If A and B are both caused by some third factor, C, then it may be that Pr(B A) > Pr(B A) even though A does not cause B. Those with yellow-stained fingers are more likely to suffer from lung cancer... smoking tends to produce both effects. Intuitively, the way to address this problem is to require that causes raise the probabilities of their effects ceteris paribus.

43 Spurious Correlations NSO Screening off: If Pr(B A C) = P(B C), then C is said to screen A off from B. Equivalently (A B) C... To avoid the problem of spurious correlations, add a no screening off (NSO) Factor A occurring at time t, is a cause of the later factor B if and only if: Pr(B A) > Pr(B A) There is no factor C, occurring earlier than or simultaneously with A, that screens A off from B.

44 Yule-Simpson Effect NSO does not suffice to resolve the problem of spurious correlations Suppose, for example, that smoking is highly correlated with exercise: those who smoke are much more likely to exercise as well. Smoking is a cause of heart disease, but suppose that exercise is an even stronger preventative of heart disease. Then it may be that smokers are, over all, less likely to suffer from heart disease than non-smokers. A smoking, C exercise, and B heart disease, Pr(B A) < Pr(B A). Note, however, that if we conditionalize on whether one exercises or not, this inequality is reversed: Pr(B A C) > Pr(B A C) Pr(B AB Mishra C) > Dobbiaco Pr(B A Lectures 2010 C). (2)Solved and Unsolved Problems in

45 Test Situations TS Causes must raise the probability of their effects in test situations: TS: A causes B if Pr(B A T) > Pr(B A T) test situation T. A test situation is a conjunction of factors, which are held fixed. This suggests that in evaluating the causal relevance of A for B, we need to hold fixed other causes of B, either positively or negatively.

46 PART V:

47 Outline

48

49 of Regulation How did a complex regulatory system like Lac operon evolve? Proximate vs. Ultimate answer... (Intelligent) Design Principles in Biology, at a systems level? Teleological Intent in Biology? Could evolution have anticipated lactose producing mammals, when it evolved E. coli? If glucose disappears in the future, will the E. coli system evolve to discard the complex regulation of diauxie?

50 Mechanisms Outline Cairns and Foster: Start with a population of E. coli lac frameshift mutants. If you starve it on medium including lactose, it appears to specifically accumulate Lac + revertants. Adaptive mutation?

51 Luria-Delbrück Jackpot 1 Causal Explanation: A stress-induced general mutagenesis mechanism in a subpopulation of starved cells causes this trait to emerge... The hypermutable state model. 2 Acausal Explanation: Stress has no direct effect on mutability but favors only growth of cells that amplify their leaky mutant lac region... The amplification mutagenesis model. Selection enhances reversion primarily by increasing the mutant lac copy number within each developing clone on the selection plate. The observed general mutagenesis is attributed to a side effect of growth with an amplificationñinduction of SOS by DNA fragments released from a tandem array of lac copies. 3 Which one is correct?

52 Acausal Explanation P[M(t + t) S(t)] = P[M(t + t) S(t)] = µ t + o( t 2 ).

53 Causal Explanation P[M(t + τ) S(t)] = ρ P[M(t + τ) S(t)] = µτ And P[M(t + t) S(t)] = µ t + o( t 2 ).

54 Statistics in the Causal System 1 Initial Population Size = N 0 ; Growth rate is β dn(t) = βn(t)dt dn(t)/dt = βdt N t = N 0 e βt. 2 R c t = (ρn 0 e βτ ) e β(t τ), t > τ. 3 Poisson Approximation E(R c t ) = ρn 0 e βt Var(R c t ) = ρn 0 e βt

55 Statistics in the Acausal System 1 Mutations occur randomly at a rate proportional to N t. Thus these mutations occur in accordance with a Poisson process whose intensity is ν(t) = µe βt. 2 The number of mutations occurring in the time interval [0, t) is M t, with expectation E(M t ) = t 0 ν(s)ds = µ β (Eβt 1).

56 1 Hence { 0, M(t) = 0; Rt a = M(t) i=1 N 0e β(t ti), o.w.. 2 Luria-Delbrück Distribution E(R a t ) = µtn 0e βt Var(R a t ) = µ β N 0e βt (e βt 1)

57 Jackpots and Fat Tails Var(Rt a) E(Rt a) = eβt 1 βt Var(Rc t ) E(Rt c) = 1.

58

59 History Outline Historically, Luria-Delbrück experiment involved pure bacterial cultures (starting with a single bacterium), attacked by a bacterial virus (bacteriophage). They noticed that the culture would clear after a few hours due to destruction of the sensitive cells by the virus. However, after further incubation for a few hours, or sometimes days, the culture would become turbid again (due to growth of a bacterial variant which is resistant to the action of the virus).

60 History Outline The observations reminded them of a casino, where a gambler would lose his bets almost always, but only very occasionally hitting a jackpot. They derived the distribution of surviving cultures (approximately) to distinguish between two hypotheses: Mutation Hypothesis vs. Acquired Hereditary Immunity Hypothesis.

61 PART VI: Time

62 Outline

63 Markov Models Suppose there are n states S 1, S 2,..., S n. And the probability of moving to a state S j from a state S i depends only on S i, but not the previous history. That is: Then by Bayes rule: P(s(t + 1) = S j s(t) = S i, s(t 1) = S i1,...) = P(s(t + 1) = S j s(t) = S i ). P(s(0) = S i0, s(1) = S i1,..., s(t 1) = S it 1, s(t) = S it ) = P(s(0) = S i0 )P(S i1 S i0 ) P(S it S it 1 ).

64 HMM: Defined with respect to an alphabet Σ States A set of (hidden) states Q, A Q Q matrix of state transition probabilities A = (a kl ), and A Q Σ matrix of emission probabilities E = (e k (σ)). Q is a set of states that emit symbols from the alphabet Σ. Dynamics is determined by a state-space trajectory determined by the state-transition probabilities.

65 A Path in the HMM Path Π = π 1 π 2 π n = a sequence of states Q in the hidden markov model, M. x Σ = sequence generated by the path Π determined by the model M: [ n ] P(x Π) = P(π 1 ) P(x i π i ) P(π i π i+1 ) i=1

66 A Path in the HMM Note that [ n ] P(x Π) = P(π 1 ) P(x i π i ) P(π i π i+1 ) P(x i π i ) = e πi (x i ) P(π i π i+1 ) = a πi,π i+1 i=1 Let π 0 and π n+1 be the initial ( begin ) and final ( end ) states, respectively P(x Π) = a π0,π 1 e π1 (x 1 )a π1,π 2 e π2 (x 2 ) e πn (x n )a πn,π n+1 i.e. n P(x Π) = a π0,π 1 e πi (x i )a πi,π i+1. i=1

67 Decoding Problem For a given sequence x, and a given path π, the model (Markovian) defines the probability P(x Π) In a casino scenario: the dealer knows Π and x, the player knows x but not Π. The path of x is hidden. Decoding Problem: Find an optimal path π for x such that P(x π) is maximized. π = arg max P(x π). π

68 Dynamic Programming Approach Principle of Optimality Optimal path for the (i + 1)-prefix of x x 1 x 2 x i+1 uses a path for an i-prefix of x that is optimal among the paths ending in an unknown state π i = k Q.

69 Dynamic Programming Approach Recurrence: s k (i) = the probability of the most probable path for the i-prefix ending in state k k Q 1 i n s k (i) = e k (x i ) max l Q s l(i 1)a lk.

70 Dynamic Programming i = 0, Base case 0 < i n, Inductive case s begin (0) = 1, s k (0) = 0, k begin. s l (i + 1) = e l (x i+1 ) max k Q [s k(i) a kl ] i = n + 1 P(x π ) = max k Q s k(n)a k,end.

71 Viterbi Algorithm Dynamic Programing with log-score function Space Complexity = O(n Q ). Time Complexity = O(n Q ). Additive formula: S l (i) = log s l (i). S l (i + 1) = log e l (x i+1 ) + max k Q [S k(i) + log a kl ].

72 [End of Lecture #2]

Bioinformatics: Biology X

Bioinformatics: Biology X Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Model Building/Checking, Reverse Engineering, Causality Outline 1 Where (or of what) one cannot speak, one must pass over in silence.

More information

Computational Systems Biology: Biology X

Computational Systems Biology: Biology X Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA L#3:(Feb-02 & 09-2010) Genome Wide Association Studies Outline 1 2 The law of causality... is a relic of a bygone age, surviving,

More information

Computational Systems Biology: Biology X

Computational Systems Biology: Biology X Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA L#7:(Mar-23-2010) Genome Wide Association Studies 1 The law of causality... is a relic of a bygone age, surviving, like the monarchy,

More information

Bioinformatics: Biology X

Bioinformatics: Biology X Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Model Building/Checking, Reverse Engineering, Causality Outline 1 Bayesian Interpretation of Probabilities 2 Where (or of what)

More information

Computational Biology Lecture #3: Probability and Statistics. Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept

Computational Biology Lecture #3: Probability and Statistics. Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept Computational Biology Lecture #3: Probability and Statistics Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept 26 2005 L2-1 Basic Probabilities L2-2 1 Random Variables L2-3 Examples

More information

Plan for today. ! Part 1: (Hidden) Markov models. ! Part 2: String matching and read mapping

Plan for today. ! Part 1: (Hidden) Markov models. ! Part 2: String matching and read mapping Plan for today! Part 1: (Hidden) Markov models! Part 2: String matching and read mapping! 2.1 Exact algorithms! 2.2 Heuristic methods for approximate search (Hidden) Markov models Why consider probabilistics

More information

11.3 Decoding Algorithm

11.3 Decoding Algorithm 11.3 Decoding Algorithm 393 For convenience, we have introduced π 0 and π n+1 as the fictitious initial and terminal states begin and end. This model defines the probability P(x π) for a given sequence

More information

An Introduction to Bioinformatics Algorithms Hidden Markov Models

An Introduction to Bioinformatics Algorithms  Hidden Markov Models Hidden Markov Models Hidden Markov Models Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm Forward-Backward Algorithm Profile HMMs HMM Parameter Estimation Viterbi training

More information

Bioinformatics: Biology X

Bioinformatics: Biology X Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Model Building/Checking, Reverse Engineering, Causality Outline 1 2 Main theses Outline There are seven main propositions in the

More information

HIDDEN MARKOV MODELS

HIDDEN MARKOV MODELS HIDDEN MARKOV MODELS Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm Forward-Backward Algorithm Profile HMMs HMM Parameter Estimation Viterbi training Baum-Welch algorithm

More information

Hidden Markov Models 1

Hidden Markov Models 1 Hidden Markov Models Dinucleotide Frequency Consider all 2-mers in a sequence {AA,AC,AG,AT,CA,CC,CG,CT,GA,GC,GG,GT,TA,TC,TG,TT} Given 4 nucleotides: each with a probability of occurrence of. 4 Thus, one

More information

An Introduction to Bioinformatics Algorithms Hidden Markov Models

An Introduction to Bioinformatics Algorithms   Hidden Markov Models Hidden Markov Models Outline 1. CG-Islands 2. The Fair Bet Casino 3. Hidden Markov Model 4. Decoding Algorithm 5. Forward-Backward Algorithm 6. Profile HMMs 7. HMM Parameter Estimation 8. Viterbi Training

More information

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem

Recall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)

More information

Hidden Markov Models

Hidden Markov Models Hidden Markov Models Outline 1. CG-Islands 2. The Fair Bet Casino 3. Hidden Markov Model 4. Decoding Algorithm 5. Forward-Backward Algorithm 6. Profile HMMs 7. HMM Parameter Estimation 8. Viterbi Training

More information

Chapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang

Chapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang Chapter 4 Dynamic Bayesian Networks 2016 Fall Jin Gu, Michael Zhang Reviews: BN Representation Basic steps for BN representations Define variables Define the preliminary relations between variables Check

More information

CS839: Probabilistic Graphical Models. Lecture 2: Directed Graphical Models. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 2: Directed Graphical Models. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 2: Directed Graphical Models Theo Rekatsinas 1 Questions Questions? Waiting list Questions on other logistics 2 Section 1 1. Intro to Bayes Nets 3 Section

More information

Hidden Markov Models

Hidden Markov Models Hidden Markov Models Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm Forward-Backward Algorithm Profile HMMs HMM Parameter Estimation Viterbi training Baum-Welch algorithm

More information

Hidden Markov Models

Hidden Markov Models Hidden Markov Models Slides revised and adapted to Bioinformática 55 Engª Biomédica/IST 2005 Ana Teresa Freitas CG-Islands Given 4 nucleotides: probability of occurrence is ~ 1/4. Thus, probability of

More information

2 : Directed GMs: Bayesian Networks

2 : Directed GMs: Bayesian Networks 10-708: Probabilistic Graphical Models, Spring 2015 2 : Directed GMs: Bayesian Networks Lecturer: Eric P. Xing Scribes: Yi Cheng, Cong Lu 1 Notation Here the notations used in this course are defined:

More information

Learning in Bayesian Networks

Learning in Bayesian Networks Learning in Bayesian Networks Florian Markowetz Max-Planck-Institute for Molecular Genetics Computational Molecular Biology Berlin Berlin: 20.06.2002 1 Overview 1. Bayesian Networks Stochastic Networks

More information

Hidden Markov Models. Three classic HMM problems

Hidden Markov Models. Three classic HMM problems An Introduction to Bioinformatics Algorithms www.bioalgorithms.info Hidden Markov Models Slides revised and adapted to Computational Biology IST 2015/2016 Ana Teresa Freitas Three classic HMM problems

More information

Example: The Dishonest Casino. Hidden Markov Models. Question # 1 Evaluation. The dishonest casino model. Question # 3 Learning. Question # 2 Decoding

Example: The Dishonest Casino. Hidden Markov Models. Question # 1 Evaluation. The dishonest casino model. Question # 3 Learning. Question # 2 Decoding Example: The Dishonest Casino Hidden Markov Models Durbin and Eddy, chapter 3 Game:. You bet $. You roll 3. Casino player rolls 4. Highest number wins $ The casino has two dice: Fair die P() = P() = P(3)

More information

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models 02-710 Computational Genomics Systems biology Putting it together: Data integration using graphical models High throughput data So far in this class we discussed several different types of high throughput

More information

2 : Directed GMs: Bayesian Networks

2 : Directed GMs: Bayesian Networks 10-708: Probabilistic Graphical Models 10-708, Spring 2017 2 : Directed GMs: Bayesian Networks Lecturer: Eric P. Xing Scribes: Jayanth Koushik, Hiroaki Hayashi, Christian Perez Topic: Directed GMs 1 Types

More information

Hidden Markov Models. Ivan Gesteira Costa Filho IZKF Research Group Bioinformatics RWTH Aachen Adapted from:

Hidden Markov Models. Ivan Gesteira Costa Filho IZKF Research Group Bioinformatics RWTH Aachen Adapted from: Hidden Markov Models Ivan Gesteira Costa Filho IZKF Research Group Bioinformatics RWTH Aachen Adapted from: www.ioalgorithms.info Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm

More information

Hidden Markov Models

Hidden Markov Models Hidden Markov Models Slides revised and adapted to Bioinformática 55 Engª Biomédica/IST 2005 Ana Teresa Freitas Forward Algorithm For Markov chains we calculate the probability of a sequence, P(x) How

More information

Motivation. Bayesian Networks in Epistemology and Philosophy of Science Lecture. Overview. Organizational Issues

Motivation. Bayesian Networks in Epistemology and Philosophy of Science Lecture. Overview. Organizational Issues Bayesian Networks in Epistemology and Philosophy of Science Lecture 1: Bayesian Networks Center for Logic and Philosophy of Science Tilburg University, The Netherlands Formal Epistemology Course Northern

More information

Lecture 4 October 18th

Lecture 4 October 18th Directed and undirected graphical models Fall 2017 Lecture 4 October 18th Lecturer: Guillaume Obozinski Scribe: In this lecture, we will assume that all random variables are discrete, to keep notations

More information

1 : Introduction. 1 Course Overview. 2 Notation. 3 Representing Multivariate Distributions : Probabilistic Graphical Models , Spring 2014

1 : Introduction. 1 Course Overview. 2 Notation. 3 Representing Multivariate Distributions : Probabilistic Graphical Models , Spring 2014 10-708: Probabilistic Graphical Models 10-708, Spring 2014 1 : Introduction Lecturer: Eric P. Xing Scribes: Daniel Silva and Calvin McCarter 1 Course Overview In this lecture we introduce the concept of

More information

CS711008Z Algorithm Design and Analysis

CS711008Z Algorithm Design and Analysis .. Lecture 6. Hidden Markov model and Viterbi s decoding algorithm Institute of Computing Technology Chinese Academy of Sciences, Beijing, China . Outline The occasionally dishonest casino: an example

More information

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008

6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, etworks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence

Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence General overview Introduction Directed acyclic graphs (DAGs) and conditional independence DAGs and causal effects

More information

Hidden Markov Models for biological sequence analysis

Hidden Markov Models for biological sequence analysis Hidden Markov Models for biological sequence analysis Master in Bioinformatics UPF 2017-2018 http://comprna.upf.edu/courses/master_agb/ Eduardo Eyras Computational Genomics Pompeu Fabra University - ICREA

More information

Probability, Entropy, and Inference / More About Inference

Probability, Entropy, and Inference / More About Inference Probability, Entropy, and Inference / More About Inference Mário S. Alvim (msalvim@dcc.ufmg.br) Information Theory DCC-UFMG (2018/02) Mário S. Alvim (msalvim@dcc.ufmg.br) Probability, Entropy, and Inference

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Dobbiaco Lectures 2010 (1)Solved and Unsolved Problems in

Dobbiaco Lectures 2010 (1)Solved and Unsolved Problems in Dobbiaco Lectures 2010 (1) Solved and Unsolved Problems in Biology Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Dobbiaco Outline 1 The Central Dogma 2 Lac Operon 3 PART I :

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

{ p if x = 1 1 p if x = 0

{ p if x = 1 1 p if x = 0 Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =

More information

COMP538: Introduction to Bayesian Networks

COMP538: Introduction to Bayesian Networks COMP538: Introduction to Bayesian Networks Lecture 9: Optimal Structure Learning Nevin L. Zhang lzhang@cse.ust.hk Department of Computer Science and Engineering Hong Kong University of Science and Technology

More information

The identification of synergism in the sufficient-component cause framework

The identification of synergism in the sufficient-component cause framework * Title Page Original Article The identification of synergism in the sufficient-component cause framework Tyler J. VanderWeele Department of Health Studies, University of Chicago James M. Robins Departments

More information

Genome 373: Hidden Markov Models II. Doug Fowler

Genome 373: Hidden Markov Models II. Doug Fowler Genome 373: Hidden Markov Models II Doug Fowler Review From Hidden Markov Models I What does a Markov model describe? Review From Hidden Markov Models I A T A Markov model describes a random process of

More information

Probability and Probability Distributions. Dr. Mohammed Alahmed

Probability and Probability Distributions. Dr. Mohammed Alahmed Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten

More information

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Probabilistic Reasoning. (Mostly using Bayesian Networks) Probabilistic Reasoning (Mostly using Bayesian Networks) Introduction: Why probabilistic reasoning? The world is not deterministic. (Usually because information is limited.) Ways of coping with uncertainty

More information

Rapid Introduction to Machine Learning/ Deep Learning

Rapid Introduction to Machine Learning/ Deep Learning Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian

More information

Abstract. Three Methods and Their Limitations. N-1 Experiments Suffice to Determine the Causal Relations Among N Variables

Abstract. Three Methods and Their Limitations. N-1 Experiments Suffice to Determine the Causal Relations Among N Variables N-1 Experiments Suffice to Determine the Causal Relations Among N Variables Frederick Eberhardt Clark Glymour 1 Richard Scheines Carnegie Mellon University Abstract By combining experimental interventions

More information

Structure learning in human causal induction

Structure learning in human causal induction Structure learning in human causal induction Joshua B. Tenenbaum & Thomas L. Griffiths Department of Psychology Stanford University, Stanford, CA 94305 jbt,gruffydd @psych.stanford.edu Abstract We use

More information

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012

Outline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012 CSE 573: Artificial Intelligence Autumn 2012 Reasoning about Uncertainty & Hidden Markov Models Daniel Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 Outline

More information

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks

Intelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2016/2017 Lesson 13 24 march 2017 Reasoning with Bayesian Networks Naïve Bayesian Systems...2 Example

More information

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty

Lecture 10: Introduction to reasoning under uncertainty. Uncertainty Lecture 10: Introduction to reasoning under uncertainty Introduction to reasoning under uncertainty Review of probability Axioms and inference Conditional probability Probability distributions COMP-424,

More information

15-381: Artificial Intelligence. Hidden Markov Models (HMMs)

15-381: Artificial Intelligence. Hidden Markov Models (HMMs) 15-381: Artificial Intelligence Hidden Markov Models (HMMs) What s wrong with Bayesian networks Bayesian networks are very useful for modeling joint distributions But they have their limitations: - Cannot

More information

Dynamic Approaches: The Hidden Markov Model

Dynamic Approaches: The Hidden Markov Model Dynamic Approaches: The Hidden Markov Model Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Inference as Message

More information

Causal Bayesian networks. Peter Antal

Causal Bayesian networks. Peter Antal Causal Bayesian networks Peter Antal antal@mit.bme.hu A.I. 4/8/2015 1 Can we represent exactly (in)dependencies by a BN? From a causal model? Suff.&nec.? Can we interpret edges as causal relations with

More information

COMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma

COMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma COMS 4771 Probabilistic Reasoning via Graphical Models Nakul Verma Last time Dimensionality Reduction Linear vs non-linear Dimensionality Reduction Principal Component Analysis (PCA) Non-linear methods

More information

Hidden Markov Models (I)

Hidden Markov Models (I) GLOBEX Bioinformatics (Summer 2015) Hidden Markov Models (I) a. The model b. The decoding: Viterbi algorithm Hidden Markov models A Markov chain of states At each state, there are a set of possible observables

More information

4.1 Notation and probability review

4.1 Notation and probability review Directed and undirected graphical models Fall 2015 Lecture 4 October 21st Lecturer: Simon Lacoste-Julien Scribe: Jaime Roquero, JieYing Wu 4.1 Notation and probability review 4.1.1 Notations Let us recall

More information

From Causality, Second edition, Contents

From Causality, Second edition, Contents From Causality, Second edition, 2009. Preface to the First Edition Preface to the Second Edition page xv xix 1 Introduction to Probabilities, Graphs, and Causal Models 1 1.1 Introduction to Probability

More information

1 The Basic Counting Principles

1 The Basic Counting Principles 1 The Basic Counting Principles The Multiplication Rule If an operation consists of k steps and the first step can be performed in n 1 ways, the second step can be performed in n ways [regardless of how

More information

Causal Inference & Reasoning with Causal Bayesian Networks

Causal Inference & Reasoning with Causal Bayesian Networks Causal Inference & Reasoning with Causal Bayesian Networks Neyman-Rubin Framework Potential Outcome Framework: for each unit k and each treatment i, there is a potential outcome on an attribute U, U ik,

More information

Topic 4: Causation and conditionals

Topic 4: Causation and conditionals Causation and conditionals http://philosophy.ucsd.edu/faculty/wuthrich/ 130 Metaphysics Fall 2012 Conditionals: If p, then q. Subjunctive and indicative conditionals Analyzing counterfactuals Closeness

More information

Introduction to Bayesian Learning. Machine Learning Fall 2018

Introduction to Bayesian Learning. Machine Learning Fall 2018 Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability

More information

Hidden Markov Models. Hosein Mohimani GHC7717

Hidden Markov Models. Hosein Mohimani GHC7717 Hidden Markov Models Hosein Mohimani GHC7717 hoseinm@andrew.cmu.edu Fair et Casino Problem Dealer flips a coin and player bets on outcome Dealer use either a fair coin (head and tail equally likely) or

More information

Appendix A Lewis s Counterfactuals

Appendix A Lewis s Counterfactuals Appendix A Lewis s Counterfactuals I will briefly describe David Lewis s possible world semantics of counterfactual conditionals (1973, p. 13). Let us understand a possible world simply as a way things

More information

Bayesian Networks Inference with Probabilistic Graphical Models

Bayesian Networks Inference with Probabilistic Graphical Models 4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning

More information

Tools for causal analysis and philosophical theories of causation. Isabelle Drouet (IHPST)

Tools for causal analysis and philosophical theories of causation. Isabelle Drouet (IHPST) Tools for causal analysis and philosophical theories of causation Isabelle Drouet (IHPST) 1 What is philosophy of causation? 1. What is causation? What characterizes these relations that we label "causal"?

More information

CMPT Machine Learning. Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th

CMPT Machine Learning. Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th CMPT 882 - Machine Learning Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th Stephen Fagan sfagan@sfu.ca Overview: Introduction - Who was Bayes? - Bayesian Statistics Versus Classical Statistics

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Uncertainty and Bayesian Networks

Uncertainty and Bayesian Networks Uncertainty and Bayesian Networks Tutorial 3 Tutorial 3 1 Outline Uncertainty Probability Syntax and Semantics for Uncertainty Inference Independence and Bayes Rule Syntax and Semantics for Bayesian Networks

More information

2. AXIOMATIC PROBABILITY

2. AXIOMATIC PROBABILITY IA Probability Lent Term 2. AXIOMATIC PROBABILITY 2. The axioms The formulation for classical probability in which all outcomes or points in the sample space are equally likely is too restrictive to develop

More information

Probabilistic Graphical Networks: Definitions and Basic Results

Probabilistic Graphical Networks: Definitions and Basic Results This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical

More information

What s an HMM? Extraction with Finite State Machines e.g. Hidden Markov Models (HMMs) Hidden Markov Models (HMMs) for Information Extraction

What s an HMM? Extraction with Finite State Machines e.g. Hidden Markov Models (HMMs) Hidden Markov Models (HMMs) for Information Extraction Hidden Markov Models (HMMs) for Information Extraction Daniel S. Weld CSE 454 Extraction with Finite State Machines e.g. Hidden Markov Models (HMMs) standard sequence model in genomics, speech, NLP, What

More information

Introduction to Hidden Markov Models for Gene Prediction ECE-S690

Introduction to Hidden Markov Models for Gene Prediction ECE-S690 Introduction to Hidden Markov Models for Gene Prediction ECE-S690 Outline Markov Models The Hidden Part How can we use this for gene prediction? Learning Models Want to recognize patterns (e.g. sequence

More information

Bayesian Networks and Markov Random Fields

Bayesian Networks and Markov Random Fields Bayesian Networks and Markov Random Fields 1 Bayesian Networks We will use capital letters for random variables and lower case letters for values of those variables. A Bayesian network is a triple V, G,

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Matrix Data: Classification: Part 2 Instructor: Yizhou Sun yzsun@ccs.neu.edu September 21, 2014 Methods to Learn Matrix Data Set Data Sequence Data Time Series Graph & Network

More information

On Likelihoodism and Intelligent Design

On Likelihoodism and Intelligent Design On Likelihoodism and Intelligent Design Sebastian Lutz Draft: 2011 02 14 Abstract Two common and plausible claims in the philosophy of science are that (i) a theory that makes no predictions is not testable

More information

CSCE 478/878 Lecture 6: Bayesian Learning

CSCE 478/878 Lecture 6: Bayesian Learning Bayesian Methods Not all hypotheses are created equal (even if they are all consistent with the training data) Outline CSCE 478/878 Lecture 6: Bayesian Learning Stephen D. Scott (Adapted from Tom Mitchell

More information

10708 Graphical Models: Homework 2

10708 Graphical Models: Homework 2 10708 Graphical Models: Homework 2 Due Monday, March 18, beginning of class Feburary 27, 2013 Instructions: There are five questions (one for extra credit) on this assignment. There is a problem involves

More information

Probability. CS 3793/5233 Artificial Intelligence Probability 1

Probability. CS 3793/5233 Artificial Intelligence Probability 1 CS 3793/5233 Artificial Intelligence 1 Motivation Motivation Random Variables Semantics Dice Example Joint Dist. Ex. Axioms Agents don t have complete knowledge about the world. Agents need to make decisions

More information

Bayesian Networks Factor Graphs the Case-Factor Algorithm and the Junction Tree Algorithm

Bayesian Networks Factor Graphs the Case-Factor Algorithm and the Junction Tree Algorithm Bayesian Networks Factor Graphs the Case-Factor Algorithm and the Junction Tree Algorithm 1 Bayesian Networks We will use capital letters for random variables and lower case letters for values of those

More information

Hidden Markov Models. x 1 x 2 x 3 x K

Hidden Markov Models. x 1 x 2 x 3 x K Hidden Markov Models 1 1 1 1 2 2 2 2 K K K K x 1 x 2 x 3 x K Viterbi, Forward, Backward VITERBI FORWARD BACKWARD Initialization: V 0 (0) = 1 V k (0) = 0, for all k > 0 Initialization: f 0 (0) = 1 f k (0)

More information

5. Conditional Distributions

5. Conditional Distributions 1 of 12 7/16/2009 5:36 AM Virtual Laboratories > 3. Distributions > 1 2 3 4 5 6 7 8 5. Conditional Distributions Basic Theory As usual, we start with a random experiment with probability measure P on an

More information

CHAPTER : Prokaryotic Genetics

CHAPTER : Prokaryotic Genetics CHAPTER 13.3 13.5: Prokaryotic Genetics 1. Most bacteria are not pathogenic. Identify several important roles they play in the ecosystem and human culture. 2. How do variations arise in bacteria considering

More information

Probabilistic Models

Probabilistic Models Bayes Nets 1 Probabilistic Models Models describe how (a portion of) the world works Models are always simplifications May not account for every variable May not account for all interactions between variables

More information

4. Conditional Probability

4. Conditional Probability 1 of 13 7/15/2009 9:25 PM Virtual Laboratories > 2. Probability Spaces > 1 2 3 4 5 6 7 4. Conditional Probability Definitions and Interpretations The Basic Definition As usual, we start with a random experiment

More information

Lecture 9: Bayesian Learning

Lecture 9: Bayesian Learning Lecture 9: Bayesian Learning Cognitive Systems II - Machine Learning Part II: Special Aspects of Concept Learning Bayes Theorem, MAL / ML hypotheses, Brute-force MAP LEARNING, MDL principle, Bayes Optimal

More information

Introduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah

Introduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah Introduction to Graphical Models Srikumar Ramalingam School of Computing University of Utah Reference Christopher M. Bishop, Pattern Recognition and Machine Learning, Jonathan S. Yedidia, William T. Freeman,

More information

Lecture 15. Probabilistic Models on Graph

Lecture 15. Probabilistic Models on Graph Lecture 15. Probabilistic Models on Graph Prof. Alan Yuille Spring 2014 1 Introduction We discuss how to define probabilistic models that use richly structured probability distributions and describe how

More information

Representation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2

Representation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2 Representation Stefano Ermon, Aditya Grover Stanford University Lecture 2 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 2 1 / 32 Learning a generative model We are given a training

More information

Template-Based Representations. Sargur Srihari

Template-Based Representations. Sargur Srihari Template-Based Representations Sargur srihari@cedar.buffalo.edu 1 Topics Variable-based vs Template-based Temporal Models Basic Assumptions Dynamic Bayesian Networks Hidden Markov Models Linear Dynamical

More information

1 Presessional Probability

1 Presessional Probability 1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional

More information

Reasoning Under Uncertainty Over Time. CS 486/686: Introduction to Artificial Intelligence

Reasoning Under Uncertainty Over Time. CS 486/686: Introduction to Artificial Intelligence Reasoning Under Uncertainty Over Time CS 486/686: Introduction to Artificial Intelligence 1 Outline Reasoning under uncertainty over time Hidden Markov Models Dynamic Bayes Nets 2 Introduction So far we

More information

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.

Recall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network. ecall from last time Lecture 3: onditional independence and graph structure onditional independencies implied by a belief network Independence maps (I-maps) Factorization theorem The Bayes ball algorithm

More information

Learning Bayesian Networks Does Not Have to Be NP-Hard

Learning Bayesian Networks Does Not Have to Be NP-Hard Learning Bayesian Networks Does Not Have to Be NP-Hard Norbert Dojer Institute of Informatics, Warsaw University, Banacha, 0-097 Warszawa, Poland dojer@mimuw.edu.pl Abstract. We propose an algorithm for

More information

Bayesian networks as causal models. Peter Antal

Bayesian networks as causal models. Peter Antal Bayesian networks as causal models Peter Antal antal@mit.bme.hu A.I. 3/20/2018 1 Can we represent exactly (in)dependencies by a BN? From a causal model? Suff.&nec.? Can we interpret edges as causal relations

More information

Hidden Markov Models. x 1 x 2 x 3 x K

Hidden Markov Models. x 1 x 2 x 3 x K Hidden Markov Models 1 1 1 1 2 2 2 2 K K K K x 1 x 2 x 3 x K HiSeq X & NextSeq Viterbi, Forward, Backward VITERBI FORWARD BACKWARD Initialization: V 0 (0) = 1 V k (0) = 0, for all k > 0 Initialization:

More information

Introduction to Probabilistic Graphical Models

Introduction to Probabilistic Graphical Models Introduction to Probabilistic Graphical Models Kyu-Baek Hwang and Byoung-Tak Zhang Biointelligence Lab School of Computer Science and Engineering Seoul National University Seoul 151-742 Korea E-mail: kbhwang@bi.snu.ac.kr

More information

CS 7180: Behavioral Modeling and Decision- making in AI

CS 7180: Behavioral Modeling and Decision- making in AI CS 7180: Behavioral Modeling and Decision- making in AI Learning Probabilistic Graphical Models Prof. Amy Sliva October 31, 2012 Hidden Markov model Stochastic system represented by three matrices N =

More information

CS6220: DATA MINING TECHNIQUES

CS6220: DATA MINING TECHNIQUES CS6220: DATA MINING TECHNIQUES Chapter 8&9: Classification: Part 3 Instructor: Yizhou Sun yzsun@ccs.neu.edu March 12, 2013 Midterm Report Grade Distribution 90-100 10 80-89 16 70-79 8 60-69 4

More information

Outline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination

Outline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination Probabilistic Graphical Models COMP 790-90 Seminar Spring 2011 The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Outline It Introduction ti Representation Bayesian network Conditional Independence Inference:

More information

Introduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah

Introduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah Introduction to Graphical Models Srikumar Ramalingam School of Computing University of Utah Reference Christopher M. Bishop, Pattern Recognition and Machine Learning, Jonathan S. Yedidia, William T. Freeman,

More information