The Ensemble of RNA Structures Example: some good structures of the RNA sequence

Size: px
Start display at page:

Download "The Ensemble of RNA Structures Example: some good structures of the RNA sequence"

Transcription

1 The Ensemble of RNA Structures Example: some good structures of the RNA sequence GGGGGUAUAGCUCAGGGGUAGAGCAUUUGACUGCAGAUCAAGAGGUCCCUGGUUCAAAUCCAGGUGCCCCCU free energy in kcal/mol (((((((..((((...))))...((((...))))(((((...)))))))))))) (((((((..((((...))))...((((.(...).))))(((((...)))))))))))) ((((((((.((((...))))(((((((((..((((...))))..)))).)))))...)))))))) ((((((((.((((...))))(((((((((..((((...))))..))).))))))...)))))))) (((((((..((((...))))...((((...))))(((((...)))))))))))) (((((((..((((...))))...(((..(...)..)))(((((...)))))))))))) ((((((((.((((...)))).((((((((..((((...))))..)))).))))...)))))))) ((((((((.((((...)))).((((((((..((((...))))..))).)))))...)))))))) ((((((((.((((...))))...((((...)))).((((...)))))))))))) ((((((...((((...))))...((((...))))(((((...))))).)))))) (((((((...(((...(((...(((...)))..)))..)))...(((((...)))))))))))) ((((((((.((((...))))((((((((...((((...))))...))).)))))...)))))))) ((((((((.((((...))))((((((((...((((...))))...)).))))))...)))))))) ((((((((.((((...))))...((((.(...).)))).((((...)))))))))))) (((((((..((((...)))).((((((...).)))))...(((((...)))))))))))) (((((((..((((...))))...(((...)))(((((...)))))))))))) ((((((...((((...))))...((((.(...).))))(((((...))))).)))))) ((((((((.((((...))))(((((((((..(((...)))..)))).)))))...)))))))) ((((((((.((((...))))(((((((((..(((...)))..))).))))))...)))))))) ((((((((.((((...))))...((((...)))).((((...)))))))))))) (((((((..((((...)))).(((((...)))))...(((((...)))))))))))) ((((((...((((...))))...((((...))))(((((...))))).)))))) The set of all nc RNA structures of an RNA sequence S is called (structure) ensemble P of S. 1c M. Möhl

2 Probability of a Structure How probable is an RNA structure P for a RNA sequence S? GOAL: define probability Pr[P S]. IDEA: Think of RNA folding as a dynamic system of structures (=states of the system). Given much time, a sequence S will form every possible structure P. For each structure there is a probability for observing it at a given time. This means: we look for a probability distribution! Requirements: probability depends on energy the lower the more probable. No additional assumptions! 2c M. Möhl

3 Distribution of States in a System Definition (Boltzmann distribution) Let X = {X 1,..., X N } denote a system of states, where state X i has energy E i. The system is Boltzmann distributed with temperature T iff Pr[X i ] = exp( βe i )/Z for Z := i exp( βe i), where β = (k B T ) 1. Remarks call exp( βe i ) Boltzmann weight of X i. broadly used in physics to describe systems of whatever Boltzmann distribution is usually assumed for the thermodynamic equilibrium (i.e. after sufficiently much time) transfer to RNA easy to see: structures=states, energies why temperature? very high temperature: all states equally probable very low temperature: only best states occur k B J/K is known as Boltzmann constant; β is called inverse temperature. 3c M. Möhl

4 What next? We assume that the structure ensemble of an RNA sequence is Boltzmann distributed. What are the benefits? (More than just probabilities of structures... ) Why is it reasonable to assume Boltzmann distribution? (Well, a physicist told me... ) How to calculate probabilities efficiently? (McCaskill s algorithm) 4c M. Möhl

5 Benefits of Assuming Boltzmann Definition Probability of a structure P for S: Pr[P S] := exp( βe(p))/z. Allows more profound weighting of structures in the ensemble. We need efficient computation of partition function Z! Even more interesting: probability of structural elements Definition Probability of a base pair (i, j) for S: Pr[(i, j) S] := P (i,j) Pr[P S] Again, we need Z (and some more). Base pair probabilities enable a new view at the structure ensemble (visually but also algorithmically!). Remark: For RNA, we have real temperature, e.g. T = 37 C, which determines β = (k B T ) 1. For calculations pay attention to physical units! 5c M. Möhl

6 An Immediate Use of Base Pair Probabilities MFE structure and base pair probability dot plot 1 of a trna GGGGGUAUAGCUCAGGGGUAGAGCAUUUGACUGCAGAUCAAGAGGUCCCUGGUUCAAAUCCAGGUGCCCCCU dot.ps U G G G G G U A U A G C U C A G G G G U A G A G C A U U U G A C U G C A G A U C A A G A G G U C C C U G G U U C A A A U C C A G G U G C C C C C U G C G C G C G C G C U G A G G U A A U G G C G U C G A G G A G U A C A UU A UCC A U A GA C U C G U UG G U C C A G C C A U UC G G A A A G 1 computed by RNAfold -p G G G G G U A U A G C U C A G G G G U A G A G C A U U U G A C U G C A G A U C A A G A G G U C C C U G G U U C A A A U C C A G G U G C C C C C U G G G G G U A U A G C U C A G G G G U A G A G C A U U U G A C U G C A G A U C A A G A G G U C C C U G G U U C A A A U C C A G G U G C C C C C U G G G G G U A U A G C U C A G G G G U A G A G C A U U U G A C U G C A G A U C A A G A G G U C C C U G G U U C A A A U C C A G G U G C C C C C U 6c M. Möhl

7 Why Do We Assume Boltzmann We will give an argument from information theory. We will show: The Boltzmann distribution makes the least number of assumptions. Formally, the B.d. is the distribution with the lowest information content/maximal (Shannon) entropy. As a consequence: without further information about our system, Boltzmann is our best choice. [ What could further information mean in a biological context? ] [ low information in distribution high information in event] 7c M. Möhl

8 What is Shannon Entropy? Information of an event = how much more you know after it happened. Assume loaded dice that always yield the number 6. you know the outcome before you throw the dice. no information content (all information in distribution) Playing lottery, it is very hard to predict the lottery numbers. high information content. This is formalized as: For some random variable X = {X 1,..., X N } the amount of information I [X i ] of some event X i X is defined as I [X i ] = log b 1 Pr[X i ] = log b Pr[X i ] Entropy is the expected amount of information for X : N H[X ] = Pr[X i ]I [X i ] i=1 8c M. Möhl

9 Shannon Entropy (Example) We toss a coin. For our coin, heads and tails show up with respective probabilities p and q (not necessarily fair). How uncertain are we about the result? p = 0.5, q = 0.5 Answer: expected H = 1 information max. uncertainty 1 H = p log b p +q log 1 b q In general, define the Shannon entropy 2 as p * log2(1/p) + q * log2(1/q) p p = 1, q = 0 H = 0 no uncertainty N H( p) := p i log b p i. i=1 2 of a probability distribution p over N states X 1... X N 9c M. Möhl

10 Formalizing Least number of assumptions Example: Assume: we have N events. Without further assumptions, we will naturally assume the uniform distribution p i = 1 N. This is the uniquely defined distribution maximizing the entropy H( p) = i p i log b p i. It is found by solving the following optimization problem: maximize the function H( p) = i p i log b p i under the side condition i p i = 1. 10

11 Formalizing Least number of assumptions Theorem: Given a system of states X 1... X N and energies E i for X i. The Boltzmann distribution is the probability distribution p that maximizes Shannon entropy H( p) = N p i log b p i under the assumption that we have a probability distribution p i = 1 i i=1 and under the assumption of known average energy < E > of the system N < E >= p i E i. i=1 11

12 Proof We show that the Boltzmann distribution is uniquely obtained by solving N maximize function H( p) = p i ln p 3 i under the side conditions i=1 C 1 ( p) = i p i 1 = 0 and C 2 ( p) = i p ie i < E > = 0 by using the method of Lagrange multipliers. 3 whether using ln or log b is equivalent for maximization 12

13 Proof Using Lagrange Multipliers Following the trick of Lagrange, find the extreme value of L( p, α, β) = H( p) αc 1 ( p) βc 2 ( p). By construction, C 1 ( p) and C 2 ( p) are partial derivatives: L( p, α, β) α L 2 ( p, α, β) β = C 1 ( p) = C 2 ( p) Thus the side conditions hold at the optimum, since there all partial derivatives are 0. 13

14 Proof (Ctd.) Partial Derivatives w.r.t p j Futhermore, we need the partial derivatives with respect to p j L( p, α, β) p j = H( p) p j α C 1( p) p j = N i=1 p i ln p i p j β C 2( p) p j α i p i 1 β p j i p ie i < E > p j = (ln p j + 1) α βe j 14

15 Proof (Ctd.) Solve Equations Finally, we need to solve the system Remarks p i E i < E > = 0 (1) i p i 1 = 0 (2) for all j (ln p j + 1) α βe j = 0 (3) Resolving (3) to p j and putting into (2) yields a distribution of the same form as the Boltzmann distribution. We won t show the dependency of β = k B T 1 and < E >. i 15

16 Proof (Ctd) Equation (3) can be rewritten to: ln p j = βe j (α + 1). Thus by exponentiation on both sides p j = exp( βe j γ) = exp( βe j), (4) exp(γ) where γ = (α + 1). By substituting (4) in (2) i p i 1 = 0 we get 1 = i exp( βe i )/ exp(γ) and thus exp(γ) = i exp( βe i ) We insert this in 4 and finally obtain p j = exp( βe j) i exp( βe i) (5) 16

17 Partition Function Recall: For probabilities, Pr[P S] = exp( βe(p))/z, we need Z. Definition For an RNA sequence S, we call Z := P nc RNA structure for S exp( βe(p)) the partition function (of the RNA ensemble P) of S. Remark Naive computation of Z: exponential, since ensemble size is exponential in S. 17

18 Excursion: Counting of Structures Computation of Z similar to counting Problem of computing the partition function is similar to counting the structures in the ensemble P. Partition function is a weighted sum, in counting we weight structures by 1. How to count nc RNA structures for S? naïve: enumerate exponential efficient: DP with decomposition a la Nussinov Example: S=CGAGC for minimal loop length m=0. 18

19 Counting of Structures: S=CGAGC C 1 G 2 A 3 G 4 C 5 C 1 G 2 A 3 G 4 C 5 19

20 Counting of Structures: S=CGAGC C 1 G 2 A 3 G 4 C 5 {.} {..,()} {...,().} {...,()..,(..)} {...,()...,(..).,.(..),...(),().()} C 1 {.} {..} {...} {...,(..)} G 2 {.} {..} {...,.()} A 3 {.} {..,()} G 4 {.} C 5 19

21 Definition (Subensemble) Subensembles Define the ij-subensemble P ij of S (for 1 i j n) as P ij := set of all nc RNA ij-substructures P of S. where: Definition (RNA Substructure) An RNA structure P of S is called ij-substructure of S iff P {i,..., j} 2. Remarks Example: see last slide, P 14 = {{}, {(1, 2)}, {(1, 4)}}, P 15 = {{}, {(1, 2)}, {(1, 4)}, {(2, 5)}, {(4, 5)}, {(1, 2), (4, 5)}} ensemble P of S: P = P 1n P ij = {{}} for j < i + m (min. loop size m) 20

22 Efficient Counting of Structures Define: C ij := P ij. ( DP-matrix C ) Computation of C ij for j i m: C ij = 1, since P ij = {{}} for j i > m: recurse! P ij consists of structures P ij 1 ( j unpaired) and structures P ik 1 P k+1j 1 {{(k, j)}} ( k, j paired ), where: combines all structures in one set with all structures in a second set. Define: P Q := {P Q P P, Q Q}. 21

23 for j i > m: P ij = P ij 1 Computation of C ij i k<j m S k,s j compl. this means for C ij : recall C ij = P ij Remarks C ij = C ij 1 + P ik 1 P k+1j 1 {{(k, j)}} i k<j m S k,s j compl. C ik 1 C k+1j 1 1 by DP: compute ensemble size C 1n in O(n 3 ) time and O(n 2 ) space. why translates to + and to? all unions were disjoint! i.e.: 1.) cases in P ij consists of... are disjoint 2.) structures combined by are disjoint 22

24 Example decompose sequence S 15 =C 1 G 2 A 3 G 4 C 5 1. subsequence C 1 G 2 A 3 G 4 and C 5 unpaired C 15 C a.) k=2. C 1, A 3 G 4, base pair (2, 5) P 15 P 11 P 34 {{(2, 5)}} C 15 C 11 C 34 1 b.) k=4. C 1 G 2 A 3, base pair (4, 5) P 15 P 13 P 54 {{(4, 5)}} C 15 C 13 C 54 1 ad 2b.) P 13 P 54 {{(4, 5)}} = {{}, {(1, 2)}} {{}} {{(4, 5)}} = {{(4, 5)}, {(1, 2), (4, 5)}} 23

25 Example decompose sequence S 15 =C 1 G 2 A 3 G 4 C 5 1. subsequence C 1 G 2 A 3 G 4 and C 5 unpaired C 15 C a.) k=2. C 1, A 3 G 4, base pair (2, 5) P 15 P 11 P 34 {{(2, 5)}} C 15 C 11 C 34 1 b.) k=4. C 1 G 2 A 3, base pair (4, 5) P 15 P 13 P 54 {{(4, 5)}} C 15 C 13 C 54 1 ad 2b.) P 13 P 54 {{(4, 5)}} = {{}, {(1, 2)}} {{}} {{(4, 5)}} = {{(4, 5)}, {(1, 2), (4, 5)}} 23

26 Counting Counting vs. Structure Prediction init C ij = 1 (j i m) recurse C ij = C ij 1 + i k<j m C ik 1 C k+1j 1 1 S k,s j compl. Prediction init N ij = 0 (j i m) recurse N ij = max{n ij 1, max i k<j m N ik 1 + N k+1j 1 + 1} S k,s j compl. Remarks translation Prediction Counting : max +, + only possible since sets disjoint, i.e. disjoint cases (no ambiguity ) non-overlapping decomposition in each single case 24

27 Back to Computing the Partition Function Recall: For probabilities, Pr[P S] = exp( βe(p))/z, we need Z. We defined: Z := P P exp( βe(p)) We claimed: Problem of computing the partition function is similar to counting the structures in the ensemble P. Partition function is a weighted sum, in counting we weight structures by 1. Definition (Partition Function of a Set of Structures) In analogy to C ij = P ij = P P ij 1, define the partition function Z P for the set of RNA structures P of S by Z P := P P exp( βe(p)). Idea: compute the Z Pij recursively efficient by DP. 25

28 Disjoint Decomposition when to add? Definition (Disjoint Sets) Two sets of RNA structures P 1 and P 2 are (structurally) disjoint iff P 1 P 2 = {}. Proposition (Disjoint Decomposition) Let P, P 1, and P 2 be sets of structures of an RNA sequence S. If P 1 and P 2 are structurally disjoint and P = P 1 P 2, then Z P = Z P1 + Z P2. 26

29 Proof Proof. Z P = exp( βe(p)) P P = disjoint exp( βe(p)) P P 1 P 2 = exp( βe(p)) + exp( βe(p)) P P 1 P P 2 = Z P1 + Z P2 27

30 Independent Decomposition when to multiply? Definition (Independent Sets) Let S be an RNA sequence. Two sets of nc RNA structures P 1 and P 2 for S are structurally independent iff for all P 1 P 1 and P 2 P 2 1. P 1 P 2 = {}. 2. each loop/secondary structure element of the RNA structure P = P 1 P 2 is either a loop of P 1 or one of P 2. Proposition (Independent Decomposition) Let P 1 and P 2 be structurally independent sets of nc RNA structures for RNA sequence S and P = P 1 P 2. Then: Z P = Z P1 Z P2 Remark: Condition (1) suffices for energy functions based on scoring base pairs (like in Nussinov). For loop-based energy models, we need (2), which implies E(P 1 P 2 ) = E(P 1 ) + E(P 2 ). 28

31 Proof. Z P = exp( βe(p)) P P = indep.(1) Proof exp( βe(p 1 P 2 )) P 1 P 1,P 2 P 2 = indep.(2) exp( β(e(p 1 ) + E(P 2 ))) P 1 P 1,P 2 P 2 = exp( βe(p 1 )) exp( βe(p 2 )) P 1 P 1 P 2 P 2 = exp( βe(p 1 )) exp( βe(p 2 )) P 1 P 1 P 2 P 2 = exp( βe(p 1 ))Z P2 P 1 P 1 = Z P1 Z P2 29

32 Adding and Multiplying of Partition Functions in the same way as for counts! Counting init C ij = 1 (j i m) recurse C ij = C ij 1 + i k<j m C ik 1 C k+1j 1 1 S k,s j compl. Partition Function (Nussinov Style) init ZP N ij = 1 (j i m) recurse ZP N ij = ZP N ij 1 + i k<j m ZP N ik 1 ZP N k+1j 1 exp( β E(basepair) ) S k,s j compl. Remarks E(basepair) : e.g. -1 or depending on S i and S j for base pair (i, j) This partitition function variant of the Nussinov algorithm can not compute the partition function for the loop-based energy model(!) 30

33 Way to RNA Partition Function Partition function adding/multiplying like in counting Attention: only for disjoint/independent sets Loop energy model Zuker: how to decompose structure space how to compute the energies (as sum of loop energies) What next? Develop recursions for partition function using real RNA energies Plan: rewrite Zuker-algo into its partition function variant What is missing? Is Zuker s decomposition of structure space disjoint? independent? 31

34 Zuker to Partition Function Variant: Example W ij = min { W ij 1 min i k<j m W ik 1 + V kj Z Pij = Z Pij 1 + i k<j m Z Pik 1 Z P kj 32

35 Zuker to Partition Function Variant: Example W ij = min { W ij 1 min i k<j m W ik 1 + V kj Z Pij = Z Pij 1 + i k<j m Z Pik 1 Z P kj? 32

36 Decomposition by Zuker { W ij 1 W ij = min min i k<j m W ik 1 + V kj { eh(i, j), min i<i V ij = min <j <j V i,j + esbi(i, j, i, j ) min i<k<j WM i+1k 1 + WM kj 1 + a { WM ij 1 + c, WM i+1j + c, V ij + b WM ij = min min i<k<j WM ik 1 + WM kj Remark Combine energy function for stacking and interior loop: ( esbi(i, j, i, j es(i, j) i = i + 1 j = j 1 ) := el(i, j, i, j ) otherwise 33

37 W ij = min Decomposition by Zuker { W ij 1 min i k<j m W ik 1 + V kj = i j i j i k j { eh(i, j), min i<i V ij = min <j <j V i,j + esbi(i, j, i, j ) min i<k<j WM i+1k 1 + WM kj 1 + a { WM ij 1 + c, WM i+1j + c, V ij + b WM ij = min min i<k<j WM ik 1 + WM kj Remark Combine energy function for stacking and interior loop: ( esbi(i, j, i, j es(i, j) i = i + 1 j = j 1 ) := el(i, j, i, j ) otherwise 33

38 Decomposition by Zuker { W ij 1 W ij = min min i k<j m W ik 1 + V kj { eh(i, j), min i<i V ij = min <j <j V i,j + esbi(i, j, i, j ) min i<k<j WM i+1k 1 + WM kj 1 + a = i j i j i i' j' j i k j WM ij = min { WM ij 1 + c, WM i+1j + c, V ij + b min i<k<j WM ik 1 + WM kj Remark Combine energy function for stacking and interior loop: ( esbi(i, j, i, j es(i, j) i = i + 1 j = j 1 ) := el(i, j, i, j ) otherwise 33

39 Decomposition by Zuker { W ij 1 W ij = min min i k<j m W ik 1 + V kj { eh(i, j), min i<i V ij = min <j <j V i,j + esbi(i, j, i, j ) min i<k<j WM i+1k 1 + WM kj 1 + a { WM ij 1 + c, WM i+1j + c, V ij + b WM ij = min min i<k<j WM ik 1 + WM kj = i j i j i j i j i k j Remark Combine energy function for stacking and interior loop: ( esbi(i, j, i, j es(i, j) i = i + 1 j = j 1 ) := el(i, j, i, j ) otherwise 33

40 Decomposition of Structure Space by Zuker = i j i j i k j = i j i j i i' j' j i k j = i j i j i j i j i k j 34

41 Ambiguity Example = i j i j i j i j i k j 35

42 Ambiguity Example = i j i j i j i j i k j j-1 j i 35

43 Ambiguity Example = i j i j i j i j i k j j i i+1 35

44 Ambiguity Example = i j i j i j i j i k j k j i 35

45 Ambiguity Example = i j i j i j i j i k j k j i 35

46 Ambiguity Example = i j i j i j i j i k j k j i 35

47 Ambiguity Example = i j i j i j i j i k j i i+1 k k k j-1 j 35

48 Discarding Ambiguity = i j i j i j i j i k j idea for unique choice: always cut at rightmost base pair = i j i k l j i k l j efficient way to do it with additional matrix: 36

49 Discarding Ambiguity = i j i j i j i j i k j idea for unique choice: always cut at rightmost base pair = i j i k l j i k l j efficient way to do it with additional matrix: = i j i k j i k j = i j i j-1 j i j 36

50 Discarding Ambiguity ambiguity in case for closed structures: = i j i j i i' j' j i k j 37

51 Discarding Ambiguity ambiguity in case for closed structures: = i j i j i i' j' j i k j = i j i j i i' j' j i k j 37

52 Unambiguous Decomposition: Summary = i j i j i k j = i j i j i i' j' j i k j = i j i k j i k j = i j i j-1 j i j 38

53 McCaskill-Algorithm: Matrices Given RNA sequence S = S 1... S n. As usual, P ij denotes the ij-subensemble of S. Matrix Q = (Q ij ) 1 i<j n Matrix Q b = (Q b ij ) 1 i<j n Matrix Q m = (Q m ij ) 1 i<j n Matrix Q m1 = (Qij m1 ) 1 i<j n 39

54 McCaskill-Algorithm: Matrices Given RNA sequence S = S 1... S n. As usual, P ij denotes the ij-subensemble of S. Matrix Q = (Q ij ) 1 i<j n Qij := ZPij Matrix Q b = (Q b ij ) 1 i<j n Q b ij := Z P b ij, where P b ij := {P P ij (i, j) P} Matrix Q m = (Q m ij ) 1 i<j n Q m ij := ZP m, where P m ij m ij := {P P ij P non-empty}, Z m includes multiloop-contributions (for unpaired bases, inner base pairs), see earlier modification E m. Matrix Q m1 = (Qij m1 where P m1 ij ) 1 i<j n Q m1 ij := Z m P, ij m1 := {P P ij k : (i, k) P k+1,..., j unpaired in P}. 39

55 Q-recursion = i j i j i k j Q ij =1 (i j m) Q ij =Q ij 1 + Q ik 1 Qkj b i k<j m 40

56 Q b -recursion = i j i j i i' j' j i k j Q b ij =0 (i j m) Qij b = exp( β eh(i, j)) + exp( β esbi(i, j, i, j )) Qi b j + i<i <j <j i<k<j m 1 Q m i+1k 1 Qm1 kj 1 exp( βa) 41

57 Q m -recursion and Q m1 -recursion = i j i k j i k j Qij m =0 (i j m) Qij m = i k<j m (Q m ik 1 + exp( β(k i)c)) Qm1 kj = i j i j-1 j i j Qij m1 =0 (i j m) Qij m1 = exp( βb) exp( β(j k)c) i+m<k j Q b ik 42

58 McCaskill Partition Function Summary Q ij =1 (i j m) Q ij =Q ij 1 + i k<j m Q b ij =0 (i j m) Q b ij = exp( β eh(i, j)) + + i<k<j m 1 Q ik 1 Q b kj i<i <j <j Q m i+1k 1 Qm1 kj 1 exp( βa) Qij m =0 (i j m) Qij m = (exp( β(k i)c) + Qik 1 m ) Qm1 i k<j m exp( β esbi(i, j, i, j )) Q b i j Qij m1 =0 (i j m) Qij m1 = exp( βb) exp( β(j k)c) i+m<k j Q b ik kj 43

59 McCaskill Remarks Partition function of the ensemble of S in Z = Q 1n Correctness due to disjoint (=unambiguous) and independent decomposition Complexity O(n 2 ) space, O(n 3 ) time (after bounding size of interior loops) Probabilities of a structure Pr[P S] = Z 1 exp( βe(p)) of a structure set Pr[P S] = Z 1 P P exp( βe(p)) of a base pair (i, j) (efficient!) Pr[(i, j) S] = Z 1 P (i,j) exp( βe(p)) (?) (depends on P ) 44

60 Base Pair Probabilities dot.ps G G G G G U A U A G C U C A G G G G U A G A G C A U U U G A C U G C A G A U C A A G A G G U C C C U G G U U C A A A U C C A G G U G C C C C C U G G G G G U A U A G C U C A G G G G U A G A G C A U U U G A C U G C A G A U C A A G A G G U C C C U G G U U C A A A U C C A G G U G C C C C C U G G G G G U A U A G C U C A G G G G U A G A G C A U U U G A C U G C A G A U C A A G A G G U C C C U G G U U C A A A U C C A G G U G C C C C C U Pr[(i, j) S] := Pr[P S] P (i,j) G G G G G U A U A G C U C A G G G G U A G A G C A U U U G A C U G C A G A U C A A G A G G U C C C U G G U U C A A A U C C A G G U G C C C C C U 45

61 McCaskill: Efficient Base Pair Probabilities Idea: Compute p kl := Pr[(k, l) S] recursively (DP!), recurse from long base pairs (outside) to small ones (inside) 1) simple case (external base pair) Definition (Probability of external base pair) p E kl := Pr[P E kl ], where P E kl := {P P P, (k, l) is external base pair in P}, where (k, l) is external base pair in P iff (k, l) P and (i, j) P : i < k < l < j. Z P E kl = Q 1k 1 Q b kl Q l+1n p E kl = Z P E kl Z = Q 1k 1Q b kl Q l+1n Q 1n 46

62 McCaskill: Efficient Base Pair Probabilities 2) general case a) (k, l) is external base pair b) (k, l) limits stacking, bulge, or interior loop closed by (i, j) c) (k, l) inner base pair of multiloop closed by (i, j) 47

63 McCaskill: Efficient Base Pair Probabilities 2) general case a) (k, l) is external base pair b) (k, l) limits stacking, bulge, or interior loop closed by (i, j) pkl SBI (i, j) := p ij Pr[loop i, j, k, l (i, j)] = p ij Z {P Pij P has loop i,j,k,l} Z P b ij exp( β esbi(i, j, k, l))qkl b = p ij Qij b. c) (k, l) inner base pair of multiloop closed by (i, j) 47

64 McC: Base Pair Probabilities Multiloop Case 2) general case c) (k, l) inner base pair of multiloop closed by (i, j) p M kl (i, j) := p ij Pr[multiloop with inner base pair (k, l) closed by (i, j) (i, j)] Three cases: position of (k, l) in the multiloop (i) (k,l) leftmost base pair (ii) (k,l) middle base pair (iii) (k,l) rightmost base pair 48

65 McC: Base Pair Probabilities Multiloop Case 2) general case c) (k, l) inner base pair of multiloop closed by (i, j) p M kl (i, j) := p ij Pr[multiloop with inner base pair (k, l) closed by (i, j) (i, j)] Three cases: position of (k, l) in the multiloop (i) (k,l) leftmost base pair Qkl b Qm l+1j 1 exp( β(a + b + (k i 1)c)) (ii) (k,l) middle base pair Qi+1k 1 m Qb kl Qm l+1j 1 exp( β(a + b)) (iii) (k,l) rightmost base pair Qi+1k 1 m Qb kl exp( β(a + b + (j l 1)c)) 48

66 McC Multiloop Case (Ctd.) Recall p M kl (i, j) := p ij Pr[multiloop with inner base pair (k, l) closed by (i, j) (i, j)] putting the three cases of (k, l) position together pkl M (i, j) = p ij[qkl b Qm l+1j 1 exp( β(a + b + (k i 1)c)) + Qi+1k 1 m Qb kl Qm l+1j 1 exp( β(a + b)) + Q m i+1k 1 Qb kl exp( β(a + b + (j l 1)c))]Q 1 ij 49

67 McCaskill Base Pair Probabilities Summary Remarks p kl = p E kl + i<k,l<j Recursive formula for p kl furnishes DP p SBI kl (i, j) + i<k,l<j p M kl Efficient calculation of all p kl in O(n 4 ) time/o(n 2 ) space (i, j) Time reduction to O(n 3 ) possible (not shown, but you learned the trick ) The algorithm by the p kl recursion is an outside algorithm; in contrast the algo for computing Z and the Q ij is inside. For getting the probabilities, we combined inside and outside. 50

68 Summary Part I Algorithms Nussinov Zuker McCaskill Common O(n 3 ) time, O(n 2 ) space non-crossing structure (= no pseudoknots ) Differences realism: base pairs free energy (loop-based) mfe ensemble 51

69 Next? Comparing RNAs 52

98 Algorithms in Bioinformatics I, WS 06, ZBIT, D. Huson, December 6, 2006

98 Algorithms in Bioinformatics I, WS 06, ZBIT, D. Huson, December 6, 2006 98 Algorithms in Bioinformatics I, WS 06, ZBIT, D. Huson, December 6, 2006 8.3.1 Simple energy minimization Maximizing the number of base pairs as described above does not lead to good structure predictions.

More information

Lecture 8: RNA folding

Lecture 8: RNA folding 1/16, Lecture 8: RNA folding Hamidreza Chitsaz Colorado State University chitsaz@cs.colostate.edu Fall 2018 September 13, 2018 2/16, Nearest Neighbor Model 3/16, McCaskill s Algorithm for MFE Structure

More information

Lecture 8: RNA folding

Lecture 8: RNA folding 1/16, Lecture 8: RNA folding Hamidreza Chitsaz Colorado State University chitsaz@cs.colostate.edu Spring 2018 February 15, 2018 2/16, Nearest Neighbor Model 3/16, McCaskill s Algorithm for MFE Structure

More information

Algorithms in Bioinformatics

Algorithms in Bioinformatics Algorithms in Bioinformatics Sami Khuri Department of Computer Science San José State University San José, California, USA khuri@cs.sjsu.edu www.cs.sjsu.edu/faculty/khuri RNA Structure Prediction Secondary

More information

RNA Structure Prediction and Comparison. RNA folding

RNA Structure Prediction and Comparison. RNA folding RNA Structure Prediction and Comparison Session 3 RNA folding Faculty of Technology robert@techfak.uni-bielefeld.de Bielefeld, WS 2013/2014 Base Pair Maximization This was the first structure prediction

More information

Combinatorial approaches to RNA folding Part II: Energy minimization via dynamic programming

Combinatorial approaches to RNA folding Part II: Energy minimization via dynamic programming ombinatorial approaches to RNA folding Part II: Energy minimization via dynamic programming Matthew Macauley Department of Mathematical Sciences lemson niversity http://www.math.clemson.edu/~macaule/ Math

More information

RNA secondary structure prediction. Farhat Habib

RNA secondary structure prediction. Farhat Habib RNA secondary structure prediction Farhat Habib RNA RNA is similar to DNA chemically. It is usually only a single strand. T(hyamine) is replaced by U(racil) Some forms of RNA can form secondary structures

More information

Computing the partition function and sampling for saturated secondary structures of RNA, with respect to the Turner energy model

Computing the partition function and sampling for saturated secondary structures of RNA, with respect to the Turner energy model Computing the partition function and sampling for saturated secondary structures of RNA, with respect to the Turner energy model J. Waldispühl 1,3 P. Clote 1,2, 1 Department of Biology, Higgins 355, Boston

More information

proteins are the basic building blocks and active players in the cell, and

proteins are the basic building blocks and active players in the cell, and 12 RN Secondary Structure Sources for this lecture: R. Durbin, S. Eddy,. Krogh und. Mitchison, Biological sequence analysis, ambridge, 1998 J. Setubal & J. Meidanis, Introduction to computational molecular

More information

Sparse RNA Folding Revisited: Space-Efficient Minimum Free Energy Prediction

Sparse RNA Folding Revisited: Space-Efficient Minimum Free Energy Prediction Sparse RNA Folding Revisited: Space-Efficient Minimum Free Energy Prediction Sebastian Will 1 and Hosna Jabbari 2 1 Bioinformatics/IZBI, University Leipzig, swill@csail.mit.edu 2 Ingenuity Lab, National

More information

BIOINFORMATICS. Fast evaluation of internal loops in RNA secondary structure prediction. Abstract. Introduction

BIOINFORMATICS. Fast evaluation of internal loops in RNA secondary structure prediction. Abstract. Introduction BIOINFORMATICS Fast evaluation of internal loops in RNA secondary structure prediction Abstract Motivation: Though not as abundant in known biological processes as proteins, RNA molecules serve as more

More information

CS681: Advanced Topics in Computational Biology

CS681: Advanced Topics in Computational Biology CS681: Advanced Topics in Computational Biology Can Alkan EA224 calkan@cs.bilkent.edu.tr Week 10 Lecture 1 http://www.cs.bilkent.edu.tr/~calkan/teaching/cs681/ RNA folding Prediction of secondary structure

More information

COMBINATORICS OF LOCALLY OPTIMAL RNA SECONDARY STRUCTURES

COMBINATORICS OF LOCALLY OPTIMAL RNA SECONDARY STRUCTURES COMBINATORICS OF LOCALLY OPTIMAL RNA SECONDARY STRUCTURES ÉRIC FUSY AND PETER CLOTE Abstract. It is a classical result of Stein and Waterman that the asymptotic number of RNA secondary structures is 1.104366

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 16. Random Variables: Distribution and Expectation

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 16. Random Variables: Distribution and Expectation CS 70 Discrete Mathematics and Probability Theory Spring 206 Rao and Walrand Note 6 Random Variables: Distribution and Expectation Example: Coin Flips Recall our setup of a probabilistic experiment as

More information

13 Comparative RNA analysis

13 Comparative RNA analysis 13 Comparative RNA analysis Sources for this lecture: R. Durbin, S. Eddy, A. Krogh und G. Mitchison, Biological sequence analysis, Cambridge, 1998 D.W. Mount. Bioinformatics: Sequences and Genome analysis,

More information

Sample Spaces, Random Variables

Sample Spaces, Random Variables Sample Spaces, Random Variables Moulinath Banerjee University of Michigan August 3, 22 Probabilities In talking about probabilities, the fundamental object is Ω, the sample space. (elements) in Ω are denoted

More information

Assignment 4: Solutions

Assignment 4: Solutions Math 340: Discrete Structures II Assignment 4: Solutions. Random Walks. Consider a random walk on an connected, non-bipartite, undirected graph G. Show that, in the long run, the walk will traverse each

More information

Supplementary Material

Supplementary Material Supplementary Material Sm-I Formal Description of the Sampling Process In the sequel, given an RNA molecule r consisting of n nucleotides, we denote the corresponding sequence fragment from position i

More information

Sparse RNA folding revisited: space efficient minimum free energy structure prediction

Sparse RNA folding revisited: space efficient minimum free energy structure prediction DOI 10.1186/s13015-016-0071-y Algorithms for Molecular Biology RESEARCH ARTICLE Sparse RNA folding revisited: space efficient minimum free energy structure prediction Sebastian Will 1* and Hosna Jabbari

More information

Discrete Mathematics and Probability Theory Fall 2013 Vazirani Note 12. Random Variables: Distribution and Expectation

Discrete Mathematics and Probability Theory Fall 2013 Vazirani Note 12. Random Variables: Distribution and Expectation CS 70 Discrete Mathematics and Probability Theory Fall 203 Vazirani Note 2 Random Variables: Distribution and Expectation We will now return once again to the question of how many heads in a typical sequence

More information

RNA-Strukturvorhersage Strukturelle Bioinformatik WS16/17

RNA-Strukturvorhersage Strukturelle Bioinformatik WS16/17 RNA-Strukturvorhersage Strukturelle Bioinformatik WS16/17 Dr. Stefan Simm, 01.11.2016 simm@bio.uni-frankfurt.de RNA secondary structures a. hairpin loop b. stem c. bulge loop d. interior loop e. multi

More information

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations EECS 70 Discrete Mathematics and Probability Theory Fall 204 Anant Sahai Note 5 Random Variables: Distributions, Independence, and Expectations In the last note, we saw how useful it is to have a way of

More information

{ p if x = 1 1 p if x = 0

{ p if x = 1 1 p if x = 0 Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =

More information

Complete Suboptimal Folding of RNA and the Stability of Secondary Structures

Complete Suboptimal Folding of RNA and the Stability of Secondary Structures Stefan Wuchty 1 Walter Fontana 1,2 Ivo L. Hofacker 1 Peter Schuster 1,2 1 Institut für Theoretische Chemie, Universität Wien, Währingerstrasse 17, A-1090 Wien, Austria Complete Suboptimal Folding of RNA

More information

RNA Basics. RNA bases A,C,G,U Canonical Base Pairs A-U G-C G-U. Bases can only pair with one other base. wobble pairing. 23 Hydrogen Bonds more stable

RNA Basics. RNA bases A,C,G,U Canonical Base Pairs A-U G-C G-U. Bases can only pair with one other base. wobble pairing. 23 Hydrogen Bonds more stable RNA STRUCTURE RNA Basics RNA bases A,C,G,U Canonical Base Pairs A-U G-C G-U wobble pairing Bases can only pair with one other base. 23 Hydrogen Bonds more stable RNA Basics transfer RNA (trna) messenger

More information

Computational Approaches for determination of Most Probable RNA Secondary Structure Using Different Thermodynamics Parameters

Computational Approaches for determination of Most Probable RNA Secondary Structure Using Different Thermodynamics Parameters Computational Approaches for determination of Most Probable RNA Secondary Structure Using Different Thermodynamics Parameters 1 Binod Kumar, Assistant Professor, Computer Sc. Dept, ISTAR, Vallabh Vidyanagar,

More information

CONTRAfold: RNA Secondary Structure Prediction without Physics-Based Models

CONTRAfold: RNA Secondary Structure Prediction without Physics-Based Models Supplementary Material for CONTRAfold: RNA Secondary Structure Prediction without Physics-Based Models Chuong B Do, Daniel A Woods, and Serafim Batzoglou Stanford University, Stanford, CA 94305, USA, {chuongdo,danwoods,serafim}@csstanfordedu,

More information

Weighted Activity Selection

Weighted Activity Selection Weighted Activity Selection Problem This problem is a generalization of the activity selection problem that we solvd with a greedy algorithm. Given a set of activities A = {[l, r ], [l, r ],..., [l n,

More information

Discrete Mathematics and Probability Theory Fall 2012 Vazirani Note 14. Random Variables: Distribution and Expectation

Discrete Mathematics and Probability Theory Fall 2012 Vazirani Note 14. Random Variables: Distribution and Expectation CS 70 Discrete Mathematics and Probability Theory Fall 202 Vazirani Note 4 Random Variables: Distribution and Expectation Random Variables Question: The homeworks of 20 students are collected in, randomly

More information

If the objects are replaced there are n choices each time yielding n r ways. n C r and in the textbook by g(n, r).

If the objects are replaced there are n choices each time yielding n r ways. n C r and in the textbook by g(n, r). Caveat: Not proof read. Corrections appreciated. Combinatorics In the following, n, n 1, r, etc. will denote non-negative integers. Rule 1 The number of ways of ordering n distinguishable objects (also

More information

Chapter Generating Functions

Chapter Generating Functions Chapter 8.1.1-8.1.2. Generating Functions Prof. Tesler Math 184A Fall 2017 Prof. Tesler Ch. 8. Generating Functions Math 184A / Fall 2017 1 / 63 Ordinary Generating Functions (OGF) Let a n (n = 0, 1,...)

More information

6.1 Main properties of Shannon entropy. Let X be a random variable taking values x in some alphabet with probabilities.

6.1 Main properties of Shannon entropy. Let X be a random variable taking values x in some alphabet with probabilities. Chapter 6 Quantum entropy There is a notion of entropy which quantifies the amount of uncertainty contained in an ensemble of Qbits. This is the von Neumann entropy that we introduce in this chapter. In

More information

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner Fundamentals CS 281A: Statistical Learning Theory Yangqing Jia Based on tutorial slides by Lester Mackey and Ariel Kleiner August, 2011 Outline 1 Probability 2 Statistics 3 Linear Algebra 4 Optimization

More information

Statistical. mechanics

Statistical. mechanics CHAPTER 15 Statistical Thermodynamics 1: The Concepts I. Introduction. A. Statistical mechanics is the bridge between microscopic and macroscopic world descriptions of nature. Statistical mechanics macroscopic

More information

1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ).

1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ). CS 70 Discrete Mathematics for CS Spring 2006 Vazirani Lecture 8 Conditional Probability A pharmaceutical company is marketing a new test for a certain medical condition. According to clinical trials,

More information

to satisfy the large number approximations, W W sys can be small.

to satisfy the large number approximations, W W sys can be small. Chapter 12. The canonical ensemble To discuss systems at constant T, we need to embed them with a diathermal wall in a heat bath. Note that only the system and bath need to be large for W tot and W bath

More information

POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS

POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS 1.1. The Rutherford-Chadwick-Ellis Experiment. About 90 years ago Ernest Rutherford and his collaborators at the Cavendish Laboratory in Cambridge conducted

More information

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman.

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman. Math 224 Fall 2017 Homework 1 Drew Armstrong Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman. Section 1.1, Exercises 4,5,6,7,9,12. Solutions to Book Problems.

More information

RNA Folding and Interaction Prediction: A Survey

RNA Folding and Interaction Prediction: A Survey RNA Folding and Interaction Prediction: A Survey Syed Ali Ahmed Graduate Center, City University of New York New York, NY November 19, 2015 Abstract The problem of computationally predicting the structure

More information

Discrete Mathematics for CS Spring 2006 Vazirani Lecture 22

Discrete Mathematics for CS Spring 2006 Vazirani Lecture 22 CS 70 Discrete Mathematics for CS Spring 2006 Vazirani Lecture 22 Random Variables and Expectation Question: The homeworks of 20 students are collected in, randomly shuffled and returned to the students.

More information

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations Linear Algebra in Computer Vision CSED441:Introduction to Computer Vision (2017F Lecture2: Basic Linear Algebra & Probability Bohyung Han CSE, POSTECH bhhan@postech.ac.kr Mathematics in vector space Linear

More information

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning Lecture 3: More on regularization. Bayesian vs maximum likelihood learning L2 and L1 regularization for linear estimators A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting

More information

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag

Decision Trees. Nicholas Ruozzi University of Texas at Dallas. Based on the slides of Vibhav Gogate and David Sontag Decision Trees Nicholas Ruozzi University of Texas at Dallas Based on the slides of Vibhav Gogate and David Sontag Supervised Learning Input: labelled training data i.e., data plus desired output Assumption:

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten

More information

Probability: Terminology and Examples Class 2, Jeremy Orloff and Jonathan Bloom

Probability: Terminology and Examples Class 2, Jeremy Orloff and Jonathan Bloom 1 Learning Goals Probability: Terminology and Examples Class 2, 18.05 Jeremy Orloff and Jonathan Bloom 1. Know the definitions of sample space, event and probability function. 2. Be able to organize a

More information

Introduction to Information Entropy Adapted from Papoulis (1991)

Introduction to Information Entropy Adapted from Papoulis (1991) Introduction to Information Entropy Adapted from Papoulis (1991) Federico Lombardo Papoulis, A., Probability, Random Variables and Stochastic Processes, 3rd edition, McGraw ill, 1991. 1 1. INTRODUCTION

More information

[S R (U 0 ɛ 1 ) S R (U 0 ɛ 2 ]. (0.1) k B

[S R (U 0 ɛ 1 ) S R (U 0 ɛ 2 ]. (0.1) k B Canonical ensemble (Two derivations) Determine the probability that a system S in contact with a reservoir 1 R to be in one particular microstate s with energy ɛ s. (If there is degeneracy we are picking

More information

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc. Chapter 14 From Randomness to Probability Copyright 2012, 2008, 2005 Pearson Education, Inc. Dealing with Random Phenomena A random phenomenon is a situation in which we know what outcomes could happen,

More information

Lecture 20 : Markov Chains

Lecture 20 : Markov Chains CSCI 3560 Probability and Computing Instructor: Bogdan Chlebus Lecture 0 : Markov Chains We consider stochastic processes. A process represents a system that evolves through incremental changes called

More information

Lecture notes for probability. Math 124

Lecture notes for probability. Math 124 Lecture notes for probability Math 124 What is probability? Probabilities are ratios, expressed as fractions, decimals, or percents, determined by considering results or outcomes of experiments whose result

More information

BIOINF 4120 Bioinforma2cs 2 - Structures and Systems -

BIOINF 4120 Bioinforma2cs 2 - Structures and Systems - BIOINF 4120 Bioinforma2cs 2 - Structures and Systems - Oliver Kohlbacher Summer 2014 3. RNA Structure Part II Overview RNA Folding Free energy as a criterion Folding free energy of RNA Zuker- SCegler algorithm

More information

MA : Introductory Probability

MA : Introductory Probability MA 320-001: Introductory Probability David Murrugarra Department of Mathematics, University of Kentucky http://www.math.uky.edu/~dmu228/ma320/ Spring 2017 David Murrugarra (University of Kentucky) MA 320:

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

Introduction to Probability

Introduction to Probability Introduction to Probability Gambling at its core 16th century Cardano: Books on Games of Chance First systematic treatment of probability 17th century Chevalier de Mere posed a problem to his friend Pascal.

More information

CS 446 Machine Learning Fall 2016 Nov 01, Bayesian Learning

CS 446 Machine Learning Fall 2016 Nov 01, Bayesian Learning CS 446 Machine Learning Fall 206 Nov 0, 206 Bayesian Learning Professor: Dan Roth Scribe: Ben Zhou, C. Cervantes Overview Bayesian Learning Naive Bayes Logistic Regression Bayesian Learning So far, we

More information

Dynamic Programming Lecture #4

Dynamic Programming Lecture #4 Dynamic Programming Lecture #4 Outline: Probability Review Probability space Conditional probability Total probability Bayes rule Independent events Conditional independence Mutual independence Probability

More information

Information. = more information was provided by the outcome in #2

Information. = more information was provided by the outcome in #2 Outline First part based very loosely on [Abramson 63]. Information theory usually formulated in terms of information channels and coding will not discuss those here.. Information 2. Entropy 3. Mutual

More information

A Gentle Tutorial on Information Theory and Learning. Roni Rosenfeld. Carnegie Mellon University

A Gentle Tutorial on Information Theory and Learning. Roni Rosenfeld. Carnegie Mellon University A Gentle Tutorial on Information Theory and Learning Roni Rosenfeld Mellon University Mellon Outline First part based very loosely on [Abramson 63]. Information theory usually formulated in terms of information

More information

A PECULIAR COIN-TOSSING MODEL

A PECULIAR COIN-TOSSING MODEL A PECULIAR COIN-TOSSING MODEL EDWARD J. GREEN 1. Coin tossing according to de Finetti A coin is drawn at random from a finite set of coins. Each coin generates an i.i.d. sequence of outcomes (heads or

More information

Notes. Relations. Introduction. Notes. Relations. Notes. Definition. Example. Slides by Christopher M. Bourke Instructor: Berthe Y.

Notes. Relations. Introduction. Notes. Relations. Notes. Definition. Example. Slides by Christopher M. Bourke Instructor: Berthe Y. Relations Slides by Christopher M. Bourke Instructor: Berthe Y. Choueiry Spring 2006 Computer Science & Engineering 235 Introduction to Discrete Mathematics Sections 7.1, 7.3 7.5 of Rosen cse235@cse.unl.edu

More information

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.)

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.) MAS 108 Probability I Notes 1 Autumn 2005 Sample space, events The general setting is: We perform an experiment which can have a number of different outcomes. The sample space is the set of all possible

More information

Lecture 7: DecisionTrees

Lecture 7: DecisionTrees Lecture 7: DecisionTrees What are decision trees? Brief interlude on information theory Decision tree construction Overfitting avoidance Regression trees COMP-652, Lecture 7 - September 28, 2009 1 Recall:

More information

The Partition Function Statistical Thermodynamics. NC State University

The Partition Function Statistical Thermodynamics. NC State University Chemistry 431 Lecture 4 The Partition Function Statistical Thermodynamics NC State University Molecular Partition Functions In general, g j is the degeneracy, ε j is the energy: = j q g e βε j We assume

More information

CSC 411 Lecture 3: Decision Trees

CSC 411 Lecture 3: Decision Trees CSC 411 Lecture 3: Decision Trees Roger Grosse, Amir-massoud Farahmand, and Juan Carrasquilla University of Toronto UofT CSC 411: 03-Decision Trees 1 / 33 Today Decision Trees Simple but powerful learning

More information

Mathematical Foundations of Computer Science Lecture Outline October 18, 2018

Mathematical Foundations of Computer Science Lecture Outline October 18, 2018 Mathematical Foundations of Computer Science Lecture Outline October 18, 2018 The Total Probability Theorem. Consider events E and F. Consider a sample point ω E. Observe that ω belongs to either F or

More information

CIS 2033 Lecture 5, Fall

CIS 2033 Lecture 5, Fall CIS 2033 Lecture 5, Fall 2016 1 Instructor: David Dobor September 13, 2016 1 Supplemental reading from Dekking s textbook: Chapter2, 3. We mentioned at the beginning of this class that calculus was a prerequisite

More information

Matrix Multiplication

Matrix Multiplication Matrix Multiplication Given two vectors a = [a 1,..., a k ] and b = [b 1,..., b k ], their inner product (or dot product) is a b = k i=1 a ib i [1, 2, 3] [ 2, 4, 6] = (1 2) + (2 4) + (3 6) = 24. We can

More information

Introduction to Quantum Computing

Introduction to Quantum Computing Introduction to Quantum Computing Petros Wallden Lecture 3: Basic Quantum Mechanics 26th September 2016 School of Informatics, University of Edinburgh Resources 1. Quantum Computation and Quantum Information

More information

Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( )

Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr( ) Theorem 1.7 [Bayes' Law]: Assume that,,, are mutually disjoint events in the sample space s.t.. Then Pr Pr = Pr Pr Pr() Pr Pr. We are given three coins and are told that two of the coins are fair and the

More information

Single Maths B: Introduction to Probability

Single Maths B: Introduction to Probability Single Maths B: Introduction to Probability Overview Lecturer Email Office Homework Webpage Dr Jonathan Cumming j.a.cumming@durham.ac.uk CM233 None! http://maths.dur.ac.uk/stats/people/jac/singleb/ 1 Introduction

More information

Chapter 4: Going from microcanonical to canonical ensemble, from energy to temperature.

Chapter 4: Going from microcanonical to canonical ensemble, from energy to temperature. Chapter 4: Going from microcanonical to canonical ensemble, from energy to temperature. All calculations in statistical mechanics can be done in the microcanonical ensemble, where all copies of the system

More information

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf

Introduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf 1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Lior Wolf 2014-15 We know that X ~ B(n,p), but we do not know p. We get a random sample from X, a

More information

An introduction to basic information theory. Hampus Wessman

An introduction to basic information theory. Hampus Wessman An introduction to basic information theory Hampus Wessman Abstract We give a short and simple introduction to basic information theory, by stripping away all the non-essentials. Theoretical bounds on

More information

Combinatorial approaches to RNA folding Part I: Basics

Combinatorial approaches to RNA folding Part I: Basics Combinatorial approaches to RNA folding Part I: Basics Matthew Macauley Department of Mathematical Sciences Clemson University http://www.math.clemson.edu/~macaule/ Math 4500, Spring 2015 M. Macauley (Clemson)

More information

1 Proof techniques. CS 224W Linear Algebra, Probability, and Proof Techniques

1 Proof techniques. CS 224W Linear Algebra, Probability, and Proof Techniques 1 Proof techniques Here we will learn to prove universal mathematical statements, like the square of any odd number is odd. It s easy enough to show that this is true in specific cases for example, 3 2

More information

Sample Space: Specify all possible outcomes from an experiment. Event: Specify a particular outcome or combination of outcomes.

Sample Space: Specify all possible outcomes from an experiment. Event: Specify a particular outcome or combination of outcomes. Chapter 2 Introduction to Probability 2.1 Probability Model Probability concerns about the chance of observing certain outcome resulting from an experiment. However, since chance is an abstraction of something

More information

Matrix Factorization and Analysis

Matrix Factorization and Analysis Chapter 7 Matrix Factorization and Analysis Matrix factorizations are an important part of the practice and analysis of signal processing. They are at the heart of many signal-processing algorithms. Their

More information

The matrix approach for abstract argumentation frameworks

The matrix approach for abstract argumentation frameworks The matrix approach for abstract argumentation frameworks Claudette CAYROL, Yuming XU IRIT Report RR- -2015-01- -FR February 2015 Abstract The matrices and the operation of dual interchange are introduced

More information

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006.

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006. Markov Chains As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006 1 Introduction A (finite) Markov chain is a process with a finite number of states (or outcomes, or

More information

the time it takes until a radioactive substance undergoes a decay

the time it takes until a radioactive substance undergoes a decay 1 Probabilities 1.1 Experiments with randomness Wewillusethetermexperimentinaverygeneralwaytorefertosomeprocess that produces a random outcome. Examples: (Ask class for some first) Here are some discrete

More information

Probabilistic Systems Analysis Spring 2018 Lecture 6. Random Variables: Probability Mass Function and Expectation

Probabilistic Systems Analysis Spring 2018 Lecture 6. Random Variables: Probability Mass Function and Expectation EE 178 Probabilistic Systems Analysis Spring 2018 Lecture 6 Random Variables: Probability Mass Function and Expectation Probability Mass Function When we introduce the basic probability model in Note 1,

More information

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com 1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com Gujarati D. Basic Econometrics, Appendix

More information

Probability. VCE Maths Methods - Unit 2 - Probability

Probability. VCE Maths Methods - Unit 2 - Probability Probability Probability Tree diagrams La ice diagrams Venn diagrams Karnough maps Probability tables Union & intersection rules Conditional probability Markov chains 1 Probability Probability is the mathematics

More information

RNA Secondary Structure Prediction

RNA Secondary Structure Prediction RN Secondary Structure Prediction Perry Hooker S 531: dvanced lgorithms Prof. Mike Rosulek University of Montana December 10, 2010 Introduction Ribonucleic acid (RN) is a macromolecule that is essential

More information

BLAST: Target frequencies and information content Dannie Durand

BLAST: Target frequencies and information content Dannie Durand Computational Genomics and Molecular Biology, Fall 2016 1 BLAST: Target frequencies and information content Dannie Durand BLAST has two components: a fast heuristic for searching for similar sequences

More information

Naive Bayes classification

Naive Bayes classification Naive Bayes classification Christos Dimitrakakis December 4, 2015 1 Introduction One of the most important methods in machine learning and statistics is that of Bayesian inference. This is the most fundamental

More information

MAT Mathematics in Today's World

MAT Mathematics in Today's World MAT 1000 Mathematics in Today's World Last Time We discussed the four rules that govern probabilities: 1. Probabilities are numbers between 0 and 1 2. The probability an event does not occur is 1 minus

More information

The integers. Chapter 3

The integers. Chapter 3 Chapter 3 The integers Recall that an abelian group is a set A with a special element 0, and operation + such that x +0=x x + y = y + x x +y + z) =x + y)+z every element x has an inverse x + y =0 We also

More information

Qualifier: CS 6375 Machine Learning Spring 2015

Qualifier: CS 6375 Machine Learning Spring 2015 Qualifier: CS 6375 Machine Learning Spring 2015 The exam is closed book. You are allowed to use two double-sided cheat sheets and a calculator. If you run out of room for an answer, use an additional sheet

More information

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events LECTURE 1 1 Introduction The first part of our adventure is a highly selective review of probability theory, focusing especially on things that are most useful in statistics. 1.1 Sample spaces and events

More information

1 Review of The Learning Setting

1 Review of The Learning Setting COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #8 Scribe: Changyan Wang February 28, 208 Review of The Learning Setting Last class, we moved beyond the PAC model: in the PAC model we

More information

CISC 889 Bioinformatics (Spring 2004) Hidden Markov Models (II)

CISC 889 Bioinformatics (Spring 2004) Hidden Markov Models (II) CISC 889 Bioinformatics (Spring 24) Hidden Markov Models (II) a. Likelihood: forward algorithm b. Decoding: Viterbi algorithm c. Model building: Baum-Welch algorithm Viterbi training Hidden Markov models

More information

Unitary evolution: this axiom governs how the state of the quantum system evolves in time.

Unitary evolution: this axiom governs how the state of the quantum system evolves in time. CS 94- Introduction Axioms Bell Inequalities /7/7 Spring 7 Lecture Why Quantum Computation? Quantum computers are the only model of computation that escape the limitations on computation imposed by the

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours

More information

Intro. Each particle has energy that we assume to be an integer E i. Any single-particle energy is equally probable for exchange, except zero, assume

Intro. Each particle has energy that we assume to be an integer E i. Any single-particle energy is equally probable for exchange, except zero, assume Intro Take N particles 5 5 5 5 5 5 Each particle has energy that we assume to be an integer E i (above, all have 5) Particle pairs can exchange energy E i! E i +1andE j! E j 1 5 4 5 6 5 5 Total number

More information

Evolutionary Tree Analysis. Overview

Evolutionary Tree Analysis. Overview CSI/BINF 5330 Evolutionary Tree Analysis Young-Rae Cho Associate Professor Department of Computer Science Baylor University Overview Backgrounds Distance-Based Evolutionary Tree Reconstruction Character-Based

More information

Introduction to Probability and Statistics (Continued)

Introduction to Probability and Statistics (Continued) Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:

More information

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability Lecture Notes 1 Basic Probability Set Theory Elements of Probability Conditional probability Sequential Calculation of Probability Total Probability and Bayes Rule Independence Counting EE 178/278A: Basic

More information

Supplementary Material

Supplementary Material Supplementary Material Sm-I Computing Inside and Outside Probabilities In order the determine all inside and outside variables for a given sequence r L r, we decided to use the SCFG G r as the basis for

More information

Lecture 1. ABC of Probability

Lecture 1. ABC of Probability Math 408 - Mathematical Statistics Lecture 1. ABC of Probability January 16, 2013 Konstantin Zuev (USC) Math 408, Lecture 1 January 16, 2013 1 / 9 Agenda Sample Spaces Realizations, Events Axioms of Probability

More information