An Introduction to Bioinformatics Algorithms Hidden Markov Models

Similar documents
HIDDEN MARKOV MODELS

Hidden Markov Models. Ivan Gesteira Costa Filho IZKF Research Group Bioinformatics RWTH Aachen Adapted from:

Hidden Markov Models

Hidden Markov Models 1

Hidden Markov Models

An Introduction to Bioinformatics Algorithms Hidden Markov Models

Hidden Markov Models. Three classic HMM problems

Hidden Markov Models

Plan for today. ! Part 1: (Hidden) Markov models. ! Part 2: String matching and read mapping

11.3 Decoding Algorithm

Hidden Markov Models

Hidden Markov Models. Hosein Mohimani GHC7717

Example: The Dishonest Casino. Hidden Markov Models. Question # 1 Evaluation. The dishonest casino model. Question # 3 Learning. Question # 2 Decoding

Hidden Markov Models for biological sequence analysis

HMM: Parameter Estimation

Hidden Markov Models for biological sequence analysis I

CISC 889 Bioinformatics (Spring 2004) Hidden Markov Models (II)

CSCE 471/871 Lecture 3: Markov Chains and

Hidden Markov Models. x 1 x 2 x 3 x K

Stephen Scott.

Lecture 9. Intro to Hidden Markov Models (finish up)

Markov Chains and Hidden Markov Models. = stochastic, generative models

Hidden Markov Models (I)

1/22/13. Example: CpG Island. Question 2: Finding CpG Islands

CSCE 478/878 Lecture 9: Hidden. Markov. Models. Stephen Scott. Introduction. Outline. Markov. Chains. Hidden Markov Models. CSCE 478/878 Lecture 9:

Hidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010

CS711008Z Algorithm Design and Analysis

Introduction to Machine Learning CMU-10701

Hidden Markov Models. By Parisa Abedi. Slides courtesy: Eric Xing

Introduction to Hidden Markov Models for Gene Prediction ECE-S690

Hidden Markov Models. x 1 x 2 x 3 x K

Bioinformatics: Biology X

Markov Chains and Hidden Markov Models. COMP 571 Luay Nakhleh, Rice University

Hidden Markov Models. Main source: Durbin et al., Biological Sequence Alignment (Cambridge, 98)

Hidden Markov Model. Ying Wu. Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208

List of Code Challenges. About the Textbook Meet the Authors... xix Meet the Development Team... xx Acknowledgments... xxi

Lecture 4: Hidden Markov Models: An Introduction to Dynamic Decision Making. November 11, 2010

Hidden Markov Models, I. Examples. Steven R. Dunbar. Toy Models. Standard Mathematical Models. Realistic Hidden Markov Models.

O 3 O 4 O 5. q 3. q 4. Transition

Hidden Markov Models. Ron Shamir, CG 08

Advanced Data Science

HMMs and biological sequence analysis

Hidden Markov Models and some applications

6.047/6.878/HST.507 Computational Biology: Genomes, Networks, Evolution. Lecture 05. Hidden Markov Models Part II

VL Algorithmen und Datenstrukturen für Bioinformatik ( ) WS15/2016 Woche 16

Chapter 4: Hidden Markov Models

Hidden Markov Models and some applications

MACHINE LEARNING 2 UGM,HMMS Lecture 7

Hidden Markov Models. based on chapters from the book Durbin, Eddy, Krogh and Mitchison Biological Sequence Analysis via Shamir s lecture notes

ROBI POLIKAR. ECE 402/504 Lecture Hidden Markov Models IGNAL PROCESSING & PATTERN RECOGNITION ROWAN UNIVERSITY

Hidden Markov Models. x 1 x 2 x 3 x N

Computational Biology Lecture #3: Probability and Statistics. Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept

Genome 373: Hidden Markov Models II. Doug Fowler

8: Hidden Markov Models

What s an HMM? Extraction with Finite State Machines e.g. Hidden Markov Models (HMMs) Hidden Markov Models (HMMs) for Information Extraction

Grundlagen der Bioinformatik, SS 09, D. Huson, June 16, S. Durbin, S. Eddy, A. Krogh and G. Mitchison, Biological Sequence

Today s Lecture: HMMs

Giri Narasimhan. CAP 5510: Introduction to Bioinformatics. ECS 254; Phone: x3748

Hidden Markov Models

COMS 4771 Probabilistic Reasoning via Graphical Models. Nakul Verma

L23: hidden Markov models

Data Mining in Bioinformatics HMM

Bio nformatics. Lecture 3. Saad Mneimneh

11 Hidden Markov Models

Grundlagen der Bioinformatik, SS 08, D. Huson, June 16, S. Durbin, S. Eddy, A. Krogh and G. Mitchison, Biological Sequence

R. Durbin, S. Eddy, A. Krogh, G. Mitchison: Biological sequence analysis. Cambridge University Press, ISBN (Chapter 3)

Hidden Markov Models (HMMs)

6 Markov Chains and Hidden Markov Models

EECS730: Introduction to Bioinformatics

Brief Introduction of Machine Learning Techniques for Content Analysis

Hidden Markov models in population genetics and evolutionary biology

Hidden Markov Methods. Algorithms and Implementation

Computational Genomics and Molecular Biology, Fall

Pattern Recognition with Hidden Markov Modells

Hidden Markov Models. music recognition. deal with variations in - pitch - timing - timbre 2

Hidden Markov Models NIKOLAY YAKOVETS

Statistical Sequence Recognition and Training: An Introduction to HMMs

HIDDEN MARKOV MODELS IN SPEECH RECOGNITION

Basic math for biology

Hidden Markov Models The three basic HMM problems (note: change in notation) Mitch Marcus CSE 391

Sequences and Information

Biology 644: Bioinformatics

Stephen Scott.

order is number of previous outputs

Hidden Markov Models. Terminology, Representation and Basic Problems

1 What is a hidden Markov model?

Introduction to Hidden Markov Models (HMMs)

Lecture 7 Sequence analysis. Hidden Markov Models

Pair Hidden Markov Models

Hidden Markov Models (HMMs) November 14, 2017

Pairwise alignment using HMMs

6.864: Lecture 5 (September 22nd, 2005) The EM Algorithm

HMM : Viterbi algorithm - a toy example

CS 136a Lecture 7 Speech Recognition Architecture: Training models with the Forward backward algorithm

Statistical Methods for NLP

The Computational Problem. We are given a sequence of DNA and we wish to know which subsequence or concatenation of subsequences constitutes a gene.

HMM : Viterbi algorithm - a toy example

Comparative Gene Finding. BMI/CS 776 Spring 2015 Colin Dewey

Phylogenetics: Likelihood

Statistical NLP: Hidden Markov Models. Updated 12/15

Transcription:

Hidden Markov Models

Hidden Markov Models

Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm Forward-Backward Algorithm Profile HMMs HMM Parameter Estimation Viterbi training Baum-Welch algorithm

CG-Islands Given 4 nucleotides: probability of occurrence is ~ 1/4. Thus, probability of occurrence of a dinucleotide is ~ 1/16. However, the frequencies of dinucleotides in DNA sequences vary widely. In particular, CG is typically underepresented (frequency of CG is typically < 1/16)

Why CG-Islands? CG is the least frequent dinucleotide because C in CG is easily methylated and has the tendency to mutate into T afterwards However, the methylation is suppressed around genes in a genome. So, CG appears at relatively high frequency within these CG islands So, finding the CG islands in a genome is an important problem

CG Islands and the Fair Bet Casino The CG islands problem can be modeled after a problem named The Fair Bet Casino The game is to flip coins, which results in only two possible outcomes: Head or Tail. The Fair coin will give Heads and Tails with same probability ½. The Biased coin will give Heads with prob. ¾.

The Fair Bet Casino (cont d) Thus, we define the probabilities: P(H F) = P(T F) = ½ P(H B) = ¾, P(T B) = ¼ The crooked dealer changes between Fair and Biased coins with probability 10%

The Fair Bet Casino Problem Input: A sequence x = x 1 x 2 x 3 x n of coin tosses made by two possible coins (F or B). Output: A sequence π = π 1 π 2 π 3 π n, with each π i being either F or B indicating that x i is the result of tossing the Fair or Biased coin respectively.

Problem Fair Bet Casino Problem Any observed outcome of coin tosses could have been generated by any sequence of states!

Problem Fair Bet Casino Problem Any observed outcome of coin tosses could have been generated by any sequence of states! Need to incorporate a way to grade different sequences differently.

Problem Fair Bet Casino Problem Any observed outcome of coin tosses could have been generated by any sequence of states! Need to incorporate a way to grade different sequences differently. Decoding Problem

Hidden Markov Model (HMM) Can be viewed as an abstract machine with k hidden states that emits symbols from an alphabet Σ. Each state has its own probability distribution, and the machine switches between states according to this probability distribution. While in a certain state, the machine makes 2 decisions: What state should I move to next? What symbol - from the alphabet Σ - should I emit?

Why Hidden? Observers can see the emitted symbols of an HMM but have no ability to know which state the HMM is currently in. Thus, the goal is to infer the most likely hidden states of an HMM based on the given sequence of emitted symbols.

HMM Parameters Σ: set of emission characters. Ex.: Σ = {H, T} for coin tossing Σ = {1, 2, 3, 4, 5, 6} for dice tossing Q: set of hidden states, each emitting symbols from Σ. Q={F,B} for coin tossing

HMM Parameters (cont d) A = (a kl ): a Q x Q matrix of probability of changing from state k to state l. a FF = 0.9 a FB = 0.1 a BF = 0.1 a BB = 0.9 E = (e k (b)): a Q x Σ matrix of probability of emitting symbol b while being in state k. e F (0) = ½ e F (1) = ½ e B (0) = ¼ e B (1) = ¾

HMM for Fair Bet Casino The Fair Bet Casino in HMM terms: Σ = {0, 1} (0 for Tails and 1 Heads) Q = {F,B} F for Fair & B for Biased coin. Transition Probabilities A *** Emission Probabilities E Fair Biased Tails(0) Heads(1 ) Fair a FF = 0.9 a FB = 0.1 Biased a BF = 0.1 a BB = 0.9 Fair e F (0) = ½ Biased e B (0) = ¼ e F (1) = ½ e B (1) = ¾

HMM for Fair Bet Casino (cont d) HMM model for the Fair Bet Casino Problem

Hidden Paths A path π = π 1 π n in the HMM is defined as a sequence of states. Consider path π = FFFBBBBBFFF and sequence x = 01011101001 Probability that x i was emitted from state π i x 0 1 0 1 1 1 0 1 0 0 1 π = F F F B B B B B F F F P(x i π i ) ½ ½ ½ ¾ ¾ ¾ ¼ ¾ ½ ½ ½ P(π i-1 à π i ) ½ 9 / 9 10 / 1 10 / 9 10 / 9 10 / 9 10 / 9 10 / 1 10 / 9 10 / 9 10 / 10 Transition probability from state π i-1 to state π i

P(x π) Calculation P(x π): Probability that sequence x was generated by the path π: n P(x π) = P(π 0 π 1 ) Π P(x i π i ) P(π i π i+1 ) i=1 = a π0, π 1 Π e πi (x i ) a π i, π i+1

P(x π) Calculation P(x π): Probability that sequence x was generated by the path π: n P(x π) = P(π 0 π 1 ) Π P(x i π i ) P(π i π i+1 ) i=1 = a π0, π 1 Π e πi (x i ) a π i, π i+1 = Π e πi+1 (x i+1 ) a π i, π i+1 if we count from i=0 instead of i=1

Decoding Problem Goal: Find an optimal hidden path of states given observations. Input: Sequence of observations x = x 1 x n generated by an HMM M(Σ, Q, A, E) Output: A path that maximizes P(x π) over all possible paths π.

Building Manhattan for Decoding Problem Andrew Viterbi used the Manhattan grid model to solve the Decoding Problem. Every choice of π = π 1 π n corresponds to a path in the graph. The only valid direction in the graph is eastward. This graph has Q 2 (n-1) edges.

Edit Graph for Decoding Problem

Decoding Problem vs. Alignment Problem Valid directions in the alignment problem. Valid directions in the decoding problem.

Decoding Problem as Finding a The Decoding Problem is reduced to finding a longest path in the directed acyclic graph (DAG) above. Notes: the length of the path is defined as the product of its edges weights, not the sum.

Decoding Problem (cont d) Every path in the graph has the probability P(x π). The Viterbi algorithm finds the path that maximizes P(x π) among all possible paths. The Viterbi algorithm runs in O(n Q 2 ) time.

Decoding Problem: weights of edges w (k, i) (l, i+1) The weight w is given by:???

Decoding Problem: weights of edges n P(x π) = Π e πi+1 (x i+1 ). a π i, π i+1 i=0 w (k, i) (l, i+1) The weight w is given by:??

Decoding Problem: weights of edges i-th term = e πi+1 (x i+1 ). a π i, π i+1 w (k, i) (l, i+1) The weight w is given by:?

Decoding Problem: weights of edges i-th term = e πi (x i ). a π i, π i+1 = e l (x i+1 ). a kl for π i =k, π i+1 =l w (k, i) (l, i+1) The weight w=e l (x i+1 ). a kl

Decoding Problem and Dynamic Programming s l,i+1 = max k Є Q {s k,i weight of edge between (k,i) and (l,i+1)}= max k Є Q {s k,i a kl e l (x i+1 ) }= e l (x i+1 ) max k Є Q {s k,i a kl }