Dobbiaco Lectures 2010 (2)Solved and Unsolved Problems in
|
|
- Augustus McKinney
- 6 years ago
- Views:
Transcription
1 Dobbiaco Lectures 2010 (2) Solved and Unsolved Problems in Biology Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Dobbiaco
2 Outline
3 PART III : Uncertainty
4 Main theses Outline A thought is a proposition with sense. 3 A proposition is a truth-function of elementary propositions. 4 The general form of a proposition is the general form of a truth function, which is: p, ξ, ξ 5 Where (or of what) one cannot speak, one must pass over in silence. Ludwig Wittgenstein, Tractatus Logico-Philosophicus, 1921.
5 Outline
6 Random Variables A (discrete) random variable is a numerical quantity that in some experiment (involving randomness) takes a value from some (discrete) set of possible values. More formally, these are measurable maps X(ω),ω Ω, from a basic probability space (Ω, F, P) ( outcomes, a sigma field of subsets of Ω and probability measure P on F ). Events...{ω Ω X(ω) = x i }... same as {X = x i } [X assumes the value x i ].
7 Few Examples Example 1: Rolling of two six-sided dice. Random Variable might be the sum of the two numbers showing on the dice. The possible values of the random variable are 2, 3,..., 12. Example 2: Occurrence of a specific word GAATTC in a genome. Random Variable might be the number of occurrence of this word in a random genome of length The possible values of the random variable are 0, 1, 2,...,
8 The Probability Distribution The probability distribution of a discrete random variable Y is the set of values that this random variable can take, together with the set of associated probabilities. Probabilities are numbers in the range between zero and one (inclusive) that always add up to one when summed over all possible values of the random variable.
9 Bernoulli Trial Outline A Bernoulli trial is a single trial with two possible outcomes: success & failure. P(success) = p and P(failure) = 1 p q. Random variable S takes the value 1 if the trial results in failure and +1 if it results in success. P S (s) = p (1+s)/2 q (1 s)/2, s = 1,+1.
10 The Binomial Distribution A Binomial random variable is the number of successes in a fixed number n of independent Bernoulli trials (with success probability = p). Random variable Y denotes the total number of successes in the n trials. ( ) n P Y (y) = p y q n y, y = 0, 1,...,n. y
11 The Poisson Distribution A random variable Y has a Poisson distribution (with parameter λ > 0) if P Y (y) = e λ λ y, y = 0, 1,... y! The Poisson distribution often arises as a limiting form of the binomial distribution.
12 Continuous Random Variables We denote a continuous random variable by X and observed value of the random variable by x. Each random variable X with range I has an associated density function f X (x) which is defined, positive for all x and integrates to one over the range I. Prob(a < X < b) = b a f X (x)dx.
13 The Normal Distribution A random variable X has a normal or Gaussian distribution if it has range (, ) and density function f X (x) = 1 2πσ e (x µ)2 2σ 2, where µ and σ > 0 are parameters of the distribution.
14 Expectation Outline For a random variable Y, and any function g(y) of Y, the expected value of g(y) is E(g(Y)) = y g(y)p Y (y), when Y is discrete; and E(g(Y)) = y g(y)f Y (y) dy, when Y is continuous. Thus, mean(y) = E(Y) = µ(y), variance(y) = E(Y 2 ) E(Y) 2 = σ 2 (Y).
15 Conditional Probabilities Suppose that A 1 and A 2 are two events such that P(A 2 ) 0. Then the conditional probability that the event A 1 occurs, given that event A 2 occurs, denoted by P(A 1 A 2 ) is given by the formula P(A 1 A 2 ) = P(A 1 A 2 ). P(A 2 )
16 Bayes Rule Outline Suppose that A 1 and A 2 are two events such that P(A 1 ) 0 and P(A 2 ) 0. Then P(A 2 A 1 ) = P(A 2)P(A 1 A 2 ). P(A 1 )
17 Bayes Nets Outline Bayes Nets or Bayesian networks are graphical representation for probabilistic relationships among a set of random variables. Given a finite set X = {X 1,..., X n } of discrete random variables where each variable X i may take values from a finite set, denoted by Val(X i ). A Bayesian network is an annotated directed acyclic graph (DAG) G that encodes a joint probability distribution over X.
18 The graph G (Bayesian Network) is defined as follows: The nodes of the graph correspond to the random variables X 1, X 2,...,X n. The links of the graph correspond to the direct influence from one variable to the other. 1 If there is a directed link from variablex i to variable X j, variable X i will be a parent of variable X j. 2 Each node is annotated with a conditional probability distribution (CPD) that represents P(X i Pa(X i )), where Pa(X i ) denotes the parents of X i in G.
19 The pair (G, CPD) encodes the joint distribution P(X 1,..., X n ). A unique joint probability distribution over X from G is factorized as: P(X 1,..., X n ) = i (P(X i Pa(X i ))).
20
21 P(X 1 = no) = 0.8 P(X 1 = yes) = 0.2 P(X 2 = absent X 1 = no) = 0.95 P(X 2 = absent X 1 = yes) = 0.75 P(X 2 = present X 1 = no) = 0.05 P(X 2 = present X 1 = yes) = 0.25 P(X 3 = absent X 1 = no) = P(X 3 = absent X 1 = yes) = P(X 3 = absent X 1 = no) = P(X 3 = absent X 1 = yes) = P(X 4 = absent X 2 = absent, P(X 4 = absent X 2 = absent, X 3 = absent) = 0.95 X 3 = present) = 0.5 P(X 4 = absent X 2 = present, P(X 4 = absent X 2 = present, X 3 = absent) = 0.9 X 3 = present) = 0.25 P(X 4 = present X 2 = absent, P(X 4 = present X 2 = absent, X 3 = absent) = 0.05 X 3 = present) = 0.5 P(X 4 = present X 2 = present, P(X 4 = present X 2 = present, X 3 = absent) = 0.1 X 3 = present) = 0.75 P(X 5 = absent X 3 = absent) = 0.98 P(X 5 = absent X 3 = present) = 0.4 P(X 5 = present X 3 = absent) = 0.02 P(X 5 = present X 3 = present) = 0.6
22 Causal Bayesian networks A causal Bayesian network of a domain is similar as the normal Bayesian network the difference is in the explanation of the links in the Bayesian networks. In the normal Bayesian networks, the links between variables can be explained as correlation or association. In a causal Bayesian network, the links mean that the parent variables will causally influence the values of the child variables.
23 The causal influence in this thesis is defined based on manipulation criteria: Manipulation Criteria Suppose there are two variables A and B in the domain; If we can manipulate the variables in the domain, set the value of variable A as a 1 or a 2 and measure its effect on variable B, then the probability distribution of variable B will change under the conditions of the different values of variable A P(B do(a = a 1 )) P(B do(a = a 2 )).
24 Bayesian network structure learning The main task in Bayesian network structure learning: Find a structure of Bayesian network that describes the observed data the best. The problem is NP-complete. Many heuristics have been proposed to learn Bayesian network structure. There are two categories of approaches for Bayesian network structure learning: The score-and-search-based approach and The constraint-based approach.
25 The score-and-search-based approach The score-and-search-based approach: The methods in this category start from an initial structure (generated randomly or from domain knowledge) and move to the neighbors with the best score in the structure space determinately or stochastically until a local maximum of the selected criteria is reached. The greedy learning process can re-start several times with different initial structures to improve the result.
26 The constraint-based approach The constraint-based approach: The methods under this category start to test the statistical significance of the pairs of variables conditioning on other variables to induce conditional independence. The pairs of variables which pass some threshold are deemed as directly connected in the Bayesian networks. The complete Bayesian network structure is constructed from the induced conditional independence and dependence information.
27 How can Bayes Nets be Causal? The causal Markov condition is a relative of Reichenbach s thesis that conditioning on common causes will render joint effects independent of other another. One can then add the assumption of faithfulness or stability as well as to assume that all underlying systems of causal laws are deterministic. Similarly, using causal minimality (or sufficiency) assumption, one may try to justify the claim that Bayesian Nets Are All There Is to Causality... Bayes nets encode information about probabilistic independencies. Causality, if it has any connection with probability, would seem to be related to probabilistic dependence. (Cartwright, N.)
28 Probability Raising Causes produce their effects; they make them happen. So, in the right kind of population we can expect that there will be a higher frequency of the effect (E) when the cause (C) is present than when it is absent; and conversely for preventatives. What kind of populations are the right kinds? Populations in which the requisite causal process sometimes operates unimpeded and its doing so is not correlated with other processes that mask the increase in probability, such as the presence of a process preventing the effect or the absence of another positive process. Cartwright
29 PART IV:
30 Outline
31
32 The law of causality... is a relic of a bygone age, surviving, like the monarchy, only because it is erroneously supposed to do no harm... Bertrand Russell, On the Notion of Cause. Proceedings of the Aristotelian Society 13: 1-26, 1913.
33 and Correlation A fallacy, known as cum hoc ergo propter hoc (Latin for with this, therefore because of this ): Correlations do not imply causation. The Probability Raising condition + Temporal Priority Condition
34 Regularity Theories (David Hume) Causes are invariably followed by their effects: We may define a cause to be an object, followed by another, and where all the objects similar to the first, are followed by objects similar to the second. Attempts to analyze causation in terms of invariable patterns of succession are referred to as regularity theories of causation. There are a number of well-known difficulties with regularity theories, and these may be used to motivate probabilistic approaches to causation.
35 Imperfect Regularities The first difficulty is that most causes are not invariably followed by their effects. Penetrance: The presence of a disease allele does not always lead to a disease phenotype. Probabilistic theories of causation: simply requires that causes raise the probability of their effects; an effect may still occur in the absence of a cause or fail to occur in its presence. Thus smoking is a cause of lung cancer, not because all smokers develop lung cancer, but because smokers are more likely to develop lung cancer than non-smokers.
36 Imperfect Regularities: INUS condition John Stuart Mill and John Mackie offer more refined accounts of the regularities that underwrite causal relations. An INUS condition: for some effect is an insufficient but non-redundant part of an unnecessary but sufficient condition. Complexity: raises problems for the epistemology of causation.
37 INUS condition Suppose, for example, that a lit match causes a forest fire. The lighting of the match, by itself, is not sufficient; many matches are lit without ensuing forest fires. The lit match is, however, a part of some constellation of conditions that are jointly sufficient for the fire. Moreover, given that this set of conditions occurred, rather than some other set sufficient for fire, the lighting of the match was necessary: fires do not occur in such circumstances when lit matches are not present. Epistasis, and gene-environment interaction.
38 Asymmetry Outline is usually asymmetric. If A causes B, then, typically, B will not also cause A. This poses a problem for regularity theories, for it seems quite plausible that if smoking is an INUS condition for lung cancer, then lung cancer will be an INUS condition for smoking. One way of enforcing the asymmetry of causation is to stipulate that causes precede their effects in time.
39 Spurious Regularities Suppose that a cause is regularly followed by two effects. For instance, a particular allele A is pleiotropic... It causes a disease trait, but also transcription of another gene B. B may be mistakenly thought to be causing the disease. B is also an INUS condition for disease state. But it s not a cause. Whenever the barometric pressure drops below a certain level, two things happen: First, the height of the column of mercury in a barometer drops. Shortly afterwards, a storm occurs. Then, it may well also be the case that whenever the column of mercury drops, there will be a storm.
40 Causes raise the probability of their effects. This can be expressed formally using the apparatus of conditional probability. Let A, B, C,... represent factors that potentially stand in causal relations. Let P be a probability function... such that Pr(A) represents the empirical probability that factor A occurs or is instantiated. Recall: P(B A) represents the conditional probability of B, given A. Pr(A B) Pr(B A) =. Pr(A)
41 If Pr(A) is 0, then the ratio in the definition of conditional probability is undefined. (There are other ways of handling this formally). A raises the probability of B is that Pr(B A) > Pr(B A). PR Axiom PR: A causes B if and only ifpr(b A) > Pr(B A).
42 Problems Outline Probability-raising is symmetric: if Pr(B A) > P(B A), then Pr(A B) > P(A B). The causal relation, however, is typically asymmetric. Probability-raising has trouble with spurious correlations. If A and B are both caused by some third factor, C, then it may be that Pr(B A) > Pr(B A) even though A does not cause B. Those with yellow-stained fingers are more likely to suffer from lung cancer... smoking tends to produce both effects. Intuitively, the way to address this problem is to require that causes raise the probabilities of their effects ceteris paribus.
43 Spurious Correlations NSO Screening off: If Pr(B A C) = P(B C), then C is said to screen A off from B. Equivalently (A B) C... To avoid the problem of spurious correlations, add a no screening off (NSO) Factor A occurring at time t, is a cause of the later factor B if and only if: Pr(B A) > Pr(B A) There is no factor C, occurring earlier than or simultaneously with A, that screens A off from B.
44 Yule-Simpson Effect NSO does not suffice to resolve the problem of spurious correlations Suppose, for example, that smoking is highly correlated with exercise: those who smoke are much more likely to exercise as well. Smoking is a cause of heart disease, but suppose that exercise is an even stronger preventative of heart disease. Then it may be that smokers are, over all, less likely to suffer from heart disease than non-smokers. A smoking, C exercise, and B heart disease, Pr(B A) < Pr(B A). Note, however, that if we conditionalize on whether one exercises or not, this inequality is reversed: Pr(B A C) > Pr(B A C) Pr(B AB Mishra C) > Dobbiaco Pr(B A Lectures 2010 C). (2)Solved and Unsolved Problems in
45 Test Situations TS Causes must raise the probability of their effects in test situations: TS: A causes B if Pr(B A T) > Pr(B A T) test situation T. A test situation is a conjunction of factors, which are held fixed. This suggests that in evaluating the causal relevance of A for B, we need to hold fixed other causes of B, either positively or negatively.
46 PART V:
47 Outline
48
49 of Regulation How did a complex regulatory system like Lac operon evolve? Proximate vs. Ultimate answer... (Intelligent) Design Principles in Biology, at a systems level? Teleological Intent in Biology? Could evolution have anticipated lactose producing mammals, when it evolved E. coli? If glucose disappears in the future, will the E. coli system evolve to discard the complex regulation of diauxie?
50 Mechanisms Outline Cairns and Foster: Start with a population of E. coli lac frameshift mutants. If you starve it on medium including lactose, it appears to specifically accumulate Lac + revertants. Adaptive mutation?
51 Luria-Delbrück Jackpot 1 Causal Explanation: A stress-induced general mutagenesis mechanism in a subpopulation of starved cells causes this trait to emerge... The hypermutable state model. 2 Acausal Explanation: Stress has no direct effect on mutability but favors only growth of cells that amplify their leaky mutant lac region... The amplification mutagenesis model. Selection enhances reversion primarily by increasing the mutant lac copy number within each developing clone on the selection plate. The observed general mutagenesis is attributed to a side effect of growth with an amplificationñinduction of SOS by DNA fragments released from a tandem array of lac copies. 3 Which one is correct?
52 Acausal Explanation P[M(t + t) S(t)] = P[M(t + t) S(t)] = µ t + o( t 2 ).
53 Causal Explanation P[M(t + τ) S(t)] = ρ P[M(t + τ) S(t)] = µτ And P[M(t + t) S(t)] = µ t + o( t 2 ).
54 Statistics in the Causal System 1 Initial Population Size = N 0 ; Growth rate is β dn(t) = βn(t)dt dn(t)/dt = βdt N t = N 0 e βt. 2 R c t = (ρn 0 e βτ ) e β(t τ), t > τ. 3 Poisson Approximation E(R c t ) = ρn 0 e βt Var(R c t ) = ρn 0 e βt
55 Statistics in the Acausal System 1 Mutations occur randomly at a rate proportional to N t. Thus these mutations occur in accordance with a Poisson process whose intensity is ν(t) = µe βt. 2 The number of mutations occurring in the time interval [0, t) is M t, with expectation E(M t ) = t 0 ν(s)ds = µ β (Eβt 1).
56 1 Hence { 0, M(t) = 0; Rt a = M(t) i=1 N 0e β(t ti), o.w.. 2 Luria-Delbrück Distribution E(R a t ) = µtn 0e βt Var(R a t ) = µ β N 0e βt (e βt 1)
57 Jackpots and Fat Tails Var(Rt a) E(Rt a) = eβt 1 βt Var(Rc t ) E(Rt c) = 1.
58
59 History Outline Historically, Luria-Delbrück experiment involved pure bacterial cultures (starting with a single bacterium), attacked by a bacterial virus (bacteriophage). They noticed that the culture would clear after a few hours due to destruction of the sensitive cells by the virus. However, after further incubation for a few hours, or sometimes days, the culture would become turbid again (due to growth of a bacterial variant which is resistant to the action of the virus).
60 History Outline The observations reminded them of a casino, where a gambler would lose his bets almost always, but only very occasionally hitting a jackpot. They derived the distribution of surviving cultures (approximately) to distinguish between two hypotheses: Mutation Hypothesis vs. Acquired Hereditary Immunity Hypothesis.
61 PART VI: Time
62 Outline
63 Markov Models Suppose there are n states S 1, S 2,..., S n. And the probability of moving to a state S j from a state S i depends only on S i, but not the previous history. That is: Then by Bayes rule: P(s(t + 1) = S j s(t) = S i, s(t 1) = S i1,...) = P(s(t + 1) = S j s(t) = S i ). P(s(0) = S i0, s(1) = S i1,..., s(t 1) = S it 1, s(t) = S it ) = P(s(0) = S i0 )P(S i1 S i0 ) P(S it S it 1 ).
64 HMM: Defined with respect to an alphabet Σ States A set of (hidden) states Q, A Q Q matrix of state transition probabilities A = (a kl ), and A Q Σ matrix of emission probabilities E = (e k (σ)). Q is a set of states that emit symbols from the alphabet Σ. Dynamics is determined by a state-space trajectory determined by the state-transition probabilities.
65 A Path in the HMM Path Π = π 1 π 2 π n = a sequence of states Q in the hidden markov model, M. x Σ = sequence generated by the path Π determined by the model M: [ n ] P(x Π) = P(π 1 ) P(x i π i ) P(π i π i+1 ) i=1
66 A Path in the HMM Note that [ n ] P(x Π) = P(π 1 ) P(x i π i ) P(π i π i+1 ) P(x i π i ) = e πi (x i ) P(π i π i+1 ) = a πi,π i+1 i=1 Let π 0 and π n+1 be the initial ( begin ) and final ( end ) states, respectively P(x Π) = a π0,π 1 e π1 (x 1 )a π1,π 2 e π2 (x 2 ) e πn (x n )a πn,π n+1 i.e. n P(x Π) = a π0,π 1 e πi (x i )a πi,π i+1. i=1
67 Decoding Problem For a given sequence x, and a given path π, the model (Markovian) defines the probability P(x Π) In a casino scenario: the dealer knows Π and x, the player knows x but not Π. The path of x is hidden. Decoding Problem: Find an optimal path π for x such that P(x π) is maximized. π = arg max P(x π). π
68 Dynamic Programming Approach Principle of Optimality Optimal path for the (i + 1)-prefix of x x 1 x 2 x i+1 uses a path for an i-prefix of x that is optimal among the paths ending in an unknown state π i = k Q.
69 Dynamic Programming Approach Recurrence: s k (i) = the probability of the most probable path for the i-prefix ending in state k k Q 1 i n s k (i) = e k (x i ) max l Q s l(i 1)a lk.
70 Dynamic Programming i = 0, Base case 0 < i n, Inductive case s begin (0) = 1, s k (0) = 0, k begin. s l (i + 1) = e l (x i+1 ) max k Q [s k(i) a kl ] i = n + 1 P(x π ) = max k Q s k(n)a k,end.
71 Viterbi Algorithm Dynamic Programing with log-score function Space Complexity = O(n Q ). Time Complexity = O(n Q ). Additive formula: S l (i) = log s l (i). S l (i + 1) = log e l (x i+1 ) + max k Q [S k(i) + log a kl ].
72 [End of Lecture #2]
Bioinformatics: Biology X
Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Model Building/Checking, Reverse Engineering, Causality Outline 1 Where (or of what) one cannot speak, one must pass over in silence.
More informationComputational Systems Biology: Biology X
Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA L#3:(Feb-02 & 09-2010) Genome Wide Association Studies Outline 1 2 The law of causality... is a relic of a bygone age, surviving,
More informationComputational Systems Biology: Biology X
Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA L#7:(Mar-23-2010) Genome Wide Association Studies 1 The law of causality... is a relic of a bygone age, surviving, like the monarchy,
More informationBioinformatics: Biology X
Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Model Building/Checking, Reverse Engineering, Causality Outline 1 Bayesian Interpretation of Probabilities 2 Where (or of what)
More informationComputational Biology Lecture #3: Probability and Statistics. Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept
Computational Biology Lecture #3: Probability and Statistics Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept 26 2005 L2-1 Basic Probabilities L2-2 1 Random Variables L2-3 Examples
More informationPlan for today. ! Part 1: (Hidden) Markov models. ! Part 2: String matching and read mapping
Plan for today! Part 1: (Hidden) Markov models! Part 2: String matching and read mapping! 2.1 Exact algorithms! 2.2 Heuristic methods for approximate search (Hidden) Markov models Why consider probabilistics
More information11.3 Decoding Algorithm
11.3 Decoding Algorithm 393 For convenience, we have introduced π 0 and π n+1 as the fictitious initial and terminal states begin and end. This model defines the probability P(x π) for a given sequence
More informationAn Introduction to Bioinformatics Algorithms Hidden Markov Models
Hidden Markov Models Hidden Markov Models Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm Forward-Backward Algorithm Profile HMMs HMM Parameter Estimation Viterbi training
More informationBioinformatics: Biology X
Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Model Building/Checking, Reverse Engineering, Causality Outline 1 2 Main theses Outline There are seven main propositions in the
More informationHIDDEN MARKOV MODELS
HIDDEN MARKOV MODELS Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm Forward-Backward Algorithm Profile HMMs HMM Parameter Estimation Viterbi training Baum-Welch algorithm
More informationHidden Markov Models 1
Hidden Markov Models Dinucleotide Frequency Consider all 2-mers in a sequence {AA,AC,AG,AT,CA,CC,CG,CT,GA,GC,GG,GT,TA,TC,TG,TT} Given 4 nucleotides: each with a probability of occurrence of. 4 Thus, one
More informationAn Introduction to Bioinformatics Algorithms Hidden Markov Models
Hidden Markov Models Outline 1. CG-Islands 2. The Fair Bet Casino 3. Hidden Markov Model 4. Decoding Algorithm 5. Forward-Backward Algorithm 6. Profile HMMs 7. HMM Parameter Estimation 8. Viterbi Training
More informationRecall from last time: Conditional probabilities. Lecture 2: Belief (Bayesian) networks. Bayes ball. Example (continued) Example: Inference problem
Recall from last time: Conditional probabilities Our probabilistic models will compute and manipulate conditional probabilities. Given two random variables X, Y, we denote by Lecture 2: Belief (Bayesian)
More informationHidden Markov Models
Hidden Markov Models Outline 1. CG-Islands 2. The Fair Bet Casino 3. Hidden Markov Model 4. Decoding Algorithm 5. Forward-Backward Algorithm 6. Profile HMMs 7. HMM Parameter Estimation 8. Viterbi Training
More informationChapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang
Chapter 4 Dynamic Bayesian Networks 2016 Fall Jin Gu, Michael Zhang Reviews: BN Representation Basic steps for BN representations Define variables Define the preliminary relations between variables Check
More informationCS839: Probabilistic Graphical Models. Lecture 2: Directed Graphical Models. Theo Rekatsinas
CS839: Probabilistic Graphical Models Lecture 2: Directed Graphical Models Theo Rekatsinas 1 Questions Questions? Waiting list Questions on other logistics 2 Section 1 1. Intro to Bayes Nets 3 Section
More informationHidden Markov Models
Hidden Markov Models Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm Forward-Backward Algorithm Profile HMMs HMM Parameter Estimation Viterbi training Baum-Welch algorithm
More informationHidden Markov Models
Hidden Markov Models Slides revised and adapted to Bioinformática 55 Engª Biomédica/IST 2005 Ana Teresa Freitas CG-Islands Given 4 nucleotides: probability of occurrence is ~ 1/4. Thus, probability of
More information2 : Directed GMs: Bayesian Networks
10-708: Probabilistic Graphical Models, Spring 2015 2 : Directed GMs: Bayesian Networks Lecturer: Eric P. Xing Scribes: Yi Cheng, Cong Lu 1 Notation Here the notations used in this course are defined:
More informationLearning in Bayesian Networks
Learning in Bayesian Networks Florian Markowetz Max-Planck-Institute for Molecular Genetics Computational Molecular Biology Berlin Berlin: 20.06.2002 1 Overview 1. Bayesian Networks Stochastic Networks
More informationHidden Markov Models. Three classic HMM problems
An Introduction to Bioinformatics Algorithms www.bioalgorithms.info Hidden Markov Models Slides revised and adapted to Computational Biology IST 2015/2016 Ana Teresa Freitas Three classic HMM problems
More informationExample: The Dishonest Casino. Hidden Markov Models. Question # 1 Evaluation. The dishonest casino model. Question # 3 Learning. Question # 2 Decoding
Example: The Dishonest Casino Hidden Markov Models Durbin and Eddy, chapter 3 Game:. You bet $. You roll 3. Casino player rolls 4. Highest number wins $ The casino has two dice: Fair die P() = P() = P(3)
More informationComputational Genomics. Systems biology. Putting it together: Data integration using graphical models
02-710 Computational Genomics Systems biology Putting it together: Data integration using graphical models High throughput data So far in this class we discussed several different types of high throughput
More information2 : Directed GMs: Bayesian Networks
10-708: Probabilistic Graphical Models 10-708, Spring 2017 2 : Directed GMs: Bayesian Networks Lecturer: Eric P. Xing Scribes: Jayanth Koushik, Hiroaki Hayashi, Christian Perez Topic: Directed GMs 1 Types
More informationHidden Markov Models. Ivan Gesteira Costa Filho IZKF Research Group Bioinformatics RWTH Aachen Adapted from:
Hidden Markov Models Ivan Gesteira Costa Filho IZKF Research Group Bioinformatics RWTH Aachen Adapted from: www.ioalgorithms.info Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm
More informationHidden Markov Models
Hidden Markov Models Slides revised and adapted to Bioinformática 55 Engª Biomédica/IST 2005 Ana Teresa Freitas Forward Algorithm For Markov chains we calculate the probability of a sequence, P(x) How
More informationMotivation. Bayesian Networks in Epistemology and Philosophy of Science Lecture. Overview. Organizational Issues
Bayesian Networks in Epistemology and Philosophy of Science Lecture 1: Bayesian Networks Center for Logic and Philosophy of Science Tilburg University, The Netherlands Formal Epistemology Course Northern
More informationLecture 4 October 18th
Directed and undirected graphical models Fall 2017 Lecture 4 October 18th Lecturer: Guillaume Obozinski Scribe: In this lecture, we will assume that all random variables are discrete, to keep notations
More information1 : Introduction. 1 Course Overview. 2 Notation. 3 Representing Multivariate Distributions : Probabilistic Graphical Models , Spring 2014
10-708: Probabilistic Graphical Models 10-708, Spring 2014 1 : Introduction Lecturer: Eric P. Xing Scribes: Daniel Silva and Calvin McCarter 1 Course Overview In this lecture we introduce the concept of
More informationCS711008Z Algorithm Design and Analysis
.. Lecture 6. Hidden Markov model and Viterbi s decoding algorithm Institute of Computing Technology Chinese Academy of Sciences, Beijing, China . Outline The occasionally dishonest casino: an example
More information6.047 / Computational Biology: Genomes, Networks, Evolution Fall 2008
MIT OpenCourseWare http://ocw.mit.edu 6.047 / 6.878 Computational Biology: Genomes, etworks, Evolution Fall 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationGraphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence
Graphical models and causality: Directed acyclic graphs (DAGs) and conditional (in)dependence General overview Introduction Directed acyclic graphs (DAGs) and conditional independence DAGs and causal effects
More informationHidden Markov Models for biological sequence analysis
Hidden Markov Models for biological sequence analysis Master in Bioinformatics UPF 2017-2018 http://comprna.upf.edu/courses/master_agb/ Eduardo Eyras Computational Genomics Pompeu Fabra University - ICREA
More informationProbability, Entropy, and Inference / More About Inference
Probability, Entropy, and Inference / More About Inference Mário S. Alvim (msalvim@dcc.ufmg.br) Information Theory DCC-UFMG (2018/02) Mário S. Alvim (msalvim@dcc.ufmg.br) Probability, Entropy, and Inference
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationDobbiaco Lectures 2010 (1)Solved and Unsolved Problems in
Dobbiaco Lectures 2010 (1) Solved and Unsolved Problems in Biology Bud Mishra Room 1002, 715 Broadway, Courant Institute, NYU, New York, USA Dobbiaco Outline 1 The Central Dogma 2 Lac Operon 3 PART I :
More informationBayesian Networks BY: MOHAMAD ALSABBAGH
Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional
More information{ p if x = 1 1 p if x = 0
Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =
More informationCOMP538: Introduction to Bayesian Networks
COMP538: Introduction to Bayesian Networks Lecture 9: Optimal Structure Learning Nevin L. Zhang lzhang@cse.ust.hk Department of Computer Science and Engineering Hong Kong University of Science and Technology
More informationThe identification of synergism in the sufficient-component cause framework
* Title Page Original Article The identification of synergism in the sufficient-component cause framework Tyler J. VanderWeele Department of Health Studies, University of Chicago James M. Robins Departments
More informationGenome 373: Hidden Markov Models II. Doug Fowler
Genome 373: Hidden Markov Models II Doug Fowler Review From Hidden Markov Models I What does a Markov model describe? Review From Hidden Markov Models I A T A Markov model describes a random process of
More informationProbability and Probability Distributions. Dr. Mohammed Alahmed
Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about
More informationDiscrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14
CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten
More informationProbabilistic Reasoning. (Mostly using Bayesian Networks)
Probabilistic Reasoning (Mostly using Bayesian Networks) Introduction: Why probabilistic reasoning? The world is not deterministic. (Usually because information is limited.) Ways of coping with uncertainty
More informationRapid Introduction to Machine Learning/ Deep Learning
Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian
More informationAbstract. Three Methods and Their Limitations. N-1 Experiments Suffice to Determine the Causal Relations Among N Variables
N-1 Experiments Suffice to Determine the Causal Relations Among N Variables Frederick Eberhardt Clark Glymour 1 Richard Scheines Carnegie Mellon University Abstract By combining experimental interventions
More informationStructure learning in human causal induction
Structure learning in human causal induction Joshua B. Tenenbaum & Thomas L. Griffiths Department of Psychology Stanford University, Stanford, CA 94305 jbt,gruffydd @psych.stanford.edu Abstract We use
More informationOutline. CSE 573: Artificial Intelligence Autumn Agent. Partial Observability. Markov Decision Process (MDP) 10/31/2012
CSE 573: Artificial Intelligence Autumn 2012 Reasoning about Uncertainty & Hidden Markov Models Daniel Weld Many slides adapted from Dan Klein, Stuart Russell, Andrew Moore & Luke Zettlemoyer 1 Outline
More informationIntelligent Systems: Reasoning and Recognition. Reasoning with Bayesian Networks
Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2016/2017 Lesson 13 24 march 2017 Reasoning with Bayesian Networks Naïve Bayesian Systems...2 Example
More informationLecture 10: Introduction to reasoning under uncertainty. Uncertainty
Lecture 10: Introduction to reasoning under uncertainty Introduction to reasoning under uncertainty Review of probability Axioms and inference Conditional probability Probability distributions COMP-424,
More information15-381: Artificial Intelligence. Hidden Markov Models (HMMs)
15-381: Artificial Intelligence Hidden Markov Models (HMMs) What s wrong with Bayesian networks Bayesian networks are very useful for modeling joint distributions But they have their limitations: - Cannot
More informationDynamic Approaches: The Hidden Markov Model
Dynamic Approaches: The Hidden Markov Model Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Inference as Message
More informationCausal Bayesian networks. Peter Antal
Causal Bayesian networks Peter Antal antal@mit.bme.hu A.I. 4/8/2015 1 Can we represent exactly (in)dependencies by a BN? From a causal model? Suff.&nec.? Can we interpret edges as causal relations with
More informationCOMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma
COMS 4771 Probabilistic Reasoning via Graphical Models Nakul Verma Last time Dimensionality Reduction Linear vs non-linear Dimensionality Reduction Principal Component Analysis (PCA) Non-linear methods
More informationHidden Markov Models (I)
GLOBEX Bioinformatics (Summer 2015) Hidden Markov Models (I) a. The model b. The decoding: Viterbi algorithm Hidden Markov models A Markov chain of states At each state, there are a set of possible observables
More information4.1 Notation and probability review
Directed and undirected graphical models Fall 2015 Lecture 4 October 21st Lecturer: Simon Lacoste-Julien Scribe: Jaime Roquero, JieYing Wu 4.1 Notation and probability review 4.1.1 Notations Let us recall
More informationFrom Causality, Second edition, Contents
From Causality, Second edition, 2009. Preface to the First Edition Preface to the Second Edition page xv xix 1 Introduction to Probabilities, Graphs, and Causal Models 1 1.1 Introduction to Probability
More information1 The Basic Counting Principles
1 The Basic Counting Principles The Multiplication Rule If an operation consists of k steps and the first step can be performed in n 1 ways, the second step can be performed in n ways [regardless of how
More informationCausal Inference & Reasoning with Causal Bayesian Networks
Causal Inference & Reasoning with Causal Bayesian Networks Neyman-Rubin Framework Potential Outcome Framework: for each unit k and each treatment i, there is a potential outcome on an attribute U, U ik,
More informationTopic 4: Causation and conditionals
Causation and conditionals http://philosophy.ucsd.edu/faculty/wuthrich/ 130 Metaphysics Fall 2012 Conditionals: If p, then q. Subjunctive and indicative conditionals Analyzing counterfactuals Closeness
More informationIntroduction to Bayesian Learning. Machine Learning Fall 2018
Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability
More informationHidden Markov Models. Hosein Mohimani GHC7717
Hidden Markov Models Hosein Mohimani GHC7717 hoseinm@andrew.cmu.edu Fair et Casino Problem Dealer flips a coin and player bets on outcome Dealer use either a fair coin (head and tail equally likely) or
More informationAppendix A Lewis s Counterfactuals
Appendix A Lewis s Counterfactuals I will briefly describe David Lewis s possible world semantics of counterfactual conditionals (1973, p. 13). Let us understand a possible world simply as a way things
More informationBayesian Networks Inference with Probabilistic Graphical Models
4190.408 2016-Spring Bayesian Networks Inference with Probabilistic Graphical Models Byoung-Tak Zhang intelligence Lab Seoul National University 4190.408 Artificial (2016-Spring) 1 Machine Learning? Learning
More informationTools for causal analysis and philosophical theories of causation. Isabelle Drouet (IHPST)
Tools for causal analysis and philosophical theories of causation Isabelle Drouet (IHPST) 1 What is philosophy of causation? 1. What is causation? What characterizes these relations that we label "causal"?
More informationCMPT Machine Learning. Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th
CMPT 882 - Machine Learning Bayesian Learning Lecture Scribe for Week 4 Jan 30th & Feb 4th Stephen Fagan sfagan@sfu.ca Overview: Introduction - Who was Bayes? - Bayesian Statistics Versus Classical Statistics
More informationBayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016
Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several
More informationUncertainty and Bayesian Networks
Uncertainty and Bayesian Networks Tutorial 3 Tutorial 3 1 Outline Uncertainty Probability Syntax and Semantics for Uncertainty Inference Independence and Bayes Rule Syntax and Semantics for Bayesian Networks
More information2. AXIOMATIC PROBABILITY
IA Probability Lent Term 2. AXIOMATIC PROBABILITY 2. The axioms The formulation for classical probability in which all outcomes or points in the sample space are equally likely is too restrictive to develop
More informationProbabilistic Graphical Networks: Definitions and Basic Results
This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical
More informationWhat s an HMM? Extraction with Finite State Machines e.g. Hidden Markov Models (HMMs) Hidden Markov Models (HMMs) for Information Extraction
Hidden Markov Models (HMMs) for Information Extraction Daniel S. Weld CSE 454 Extraction with Finite State Machines e.g. Hidden Markov Models (HMMs) standard sequence model in genomics, speech, NLP, What
More informationIntroduction to Hidden Markov Models for Gene Prediction ECE-S690
Introduction to Hidden Markov Models for Gene Prediction ECE-S690 Outline Markov Models The Hidden Part How can we use this for gene prediction? Learning Models Want to recognize patterns (e.g. sequence
More informationBayesian Networks and Markov Random Fields
Bayesian Networks and Markov Random Fields 1 Bayesian Networks We will use capital letters for random variables and lower case letters for values of those variables. A Bayesian network is a triple V, G,
More informationCS6220: DATA MINING TECHNIQUES
CS6220: DATA MINING TECHNIQUES Matrix Data: Classification: Part 2 Instructor: Yizhou Sun yzsun@ccs.neu.edu September 21, 2014 Methods to Learn Matrix Data Set Data Sequence Data Time Series Graph & Network
More informationOn Likelihoodism and Intelligent Design
On Likelihoodism and Intelligent Design Sebastian Lutz Draft: 2011 02 14 Abstract Two common and plausible claims in the philosophy of science are that (i) a theory that makes no predictions is not testable
More informationCSCE 478/878 Lecture 6: Bayesian Learning
Bayesian Methods Not all hypotheses are created equal (even if they are all consistent with the training data) Outline CSCE 478/878 Lecture 6: Bayesian Learning Stephen D. Scott (Adapted from Tom Mitchell
More information10708 Graphical Models: Homework 2
10708 Graphical Models: Homework 2 Due Monday, March 18, beginning of class Feburary 27, 2013 Instructions: There are five questions (one for extra credit) on this assignment. There is a problem involves
More informationProbability. CS 3793/5233 Artificial Intelligence Probability 1
CS 3793/5233 Artificial Intelligence 1 Motivation Motivation Random Variables Semantics Dice Example Joint Dist. Ex. Axioms Agents don t have complete knowledge about the world. Agents need to make decisions
More informationBayesian Networks Factor Graphs the Case-Factor Algorithm and the Junction Tree Algorithm
Bayesian Networks Factor Graphs the Case-Factor Algorithm and the Junction Tree Algorithm 1 Bayesian Networks We will use capital letters for random variables and lower case letters for values of those
More informationHidden Markov Models. x 1 x 2 x 3 x K
Hidden Markov Models 1 1 1 1 2 2 2 2 K K K K x 1 x 2 x 3 x K Viterbi, Forward, Backward VITERBI FORWARD BACKWARD Initialization: V 0 (0) = 1 V k (0) = 0, for all k > 0 Initialization: f 0 (0) = 1 f k (0)
More information5. Conditional Distributions
1 of 12 7/16/2009 5:36 AM Virtual Laboratories > 3. Distributions > 1 2 3 4 5 6 7 8 5. Conditional Distributions Basic Theory As usual, we start with a random experiment with probability measure P on an
More informationCHAPTER : Prokaryotic Genetics
CHAPTER 13.3 13.5: Prokaryotic Genetics 1. Most bacteria are not pathogenic. Identify several important roles they play in the ecosystem and human culture. 2. How do variations arise in bacteria considering
More informationProbabilistic Models
Bayes Nets 1 Probabilistic Models Models describe how (a portion of) the world works Models are always simplifications May not account for every variable May not account for all interactions between variables
More information4. Conditional Probability
1 of 13 7/15/2009 9:25 PM Virtual Laboratories > 2. Probability Spaces > 1 2 3 4 5 6 7 4. Conditional Probability Definitions and Interpretations The Basic Definition As usual, we start with a random experiment
More informationLecture 9: Bayesian Learning
Lecture 9: Bayesian Learning Cognitive Systems II - Machine Learning Part II: Special Aspects of Concept Learning Bayes Theorem, MAL / ML hypotheses, Brute-force MAP LEARNING, MDL principle, Bayes Optimal
More informationIntroduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah
Introduction to Graphical Models Srikumar Ramalingam School of Computing University of Utah Reference Christopher M. Bishop, Pattern Recognition and Machine Learning, Jonathan S. Yedidia, William T. Freeman,
More informationLecture 15. Probabilistic Models on Graph
Lecture 15. Probabilistic Models on Graph Prof. Alan Yuille Spring 2014 1 Introduction We discuss how to define probabilistic models that use richly structured probability distributions and describe how
More informationRepresentation. Stefano Ermon, Aditya Grover. Stanford University. Lecture 2
Representation Stefano Ermon, Aditya Grover Stanford University Lecture 2 Stefano Ermon, Aditya Grover (AI Lab) Deep Generative Models Lecture 2 1 / 32 Learning a generative model We are given a training
More informationTemplate-Based Representations. Sargur Srihari
Template-Based Representations Sargur srihari@cedar.buffalo.edu 1 Topics Variable-based vs Template-based Temporal Models Basic Assumptions Dynamic Bayesian Networks Hidden Markov Models Linear Dynamical
More information1 Presessional Probability
1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional
More informationReasoning Under Uncertainty Over Time. CS 486/686: Introduction to Artificial Intelligence
Reasoning Under Uncertainty Over Time CS 486/686: Introduction to Artificial Intelligence 1 Outline Reasoning under uncertainty over time Hidden Markov Models Dynamic Bayes Nets 2 Introduction So far we
More informationRecall from last time. Lecture 3: Conditional independence and graph structure. Example: A Bayesian (belief) network.
ecall from last time Lecture 3: onditional independence and graph structure onditional independencies implied by a belief network Independence maps (I-maps) Factorization theorem The Bayes ball algorithm
More informationLearning Bayesian Networks Does Not Have to Be NP-Hard
Learning Bayesian Networks Does Not Have to Be NP-Hard Norbert Dojer Institute of Informatics, Warsaw University, Banacha, 0-097 Warszawa, Poland dojer@mimuw.edu.pl Abstract. We propose an algorithm for
More informationBayesian networks as causal models. Peter Antal
Bayesian networks as causal models Peter Antal antal@mit.bme.hu A.I. 3/20/2018 1 Can we represent exactly (in)dependencies by a BN? From a causal model? Suff.&nec.? Can we interpret edges as causal relations
More informationHidden Markov Models. x 1 x 2 x 3 x K
Hidden Markov Models 1 1 1 1 2 2 2 2 K K K K x 1 x 2 x 3 x K HiSeq X & NextSeq Viterbi, Forward, Backward VITERBI FORWARD BACKWARD Initialization: V 0 (0) = 1 V k (0) = 0, for all k > 0 Initialization:
More informationIntroduction to Probabilistic Graphical Models
Introduction to Probabilistic Graphical Models Kyu-Baek Hwang and Byoung-Tak Zhang Biointelligence Lab School of Computer Science and Engineering Seoul National University Seoul 151-742 Korea E-mail: kbhwang@bi.snu.ac.kr
More informationCS 7180: Behavioral Modeling and Decision- making in AI
CS 7180: Behavioral Modeling and Decision- making in AI Learning Probabilistic Graphical Models Prof. Amy Sliva October 31, 2012 Hidden Markov model Stochastic system represented by three matrices N =
More informationCS6220: DATA MINING TECHNIQUES
CS6220: DATA MINING TECHNIQUES Chapter 8&9: Classification: Part 3 Instructor: Yizhou Sun yzsun@ccs.neu.edu March 12, 2013 Midterm Report Grade Distribution 90-100 10 80-89 16 70-79 8 60-69 4
More informationOutline. Spring It Introduction Representation. Markov Random Field. Conclusion. Conditional Independence Inference: Variable elimination
Probabilistic Graphical Models COMP 790-90 Seminar Spring 2011 The UNIVERSITY of NORTH CAROLINA at CHAPEL HILL Outline It Introduction ti Representation Bayesian network Conditional Independence Inference:
More informationIntroduction to Graphical Models. Srikumar Ramalingam School of Computing University of Utah
Introduction to Graphical Models Srikumar Ramalingam School of Computing University of Utah Reference Christopher M. Bishop, Pattern Recognition and Machine Learning, Jonathan S. Yedidia, William T. Freeman,
More information