Advanced Algorithms. Lecture Notes for April 5, 2016 Dynamic programming, continued (HMMs); Iterative improvement Bernard Moret
|
|
- Edwina Ward
- 5 years ago
- Views:
Transcription
1 Advanced Algorithms Lecture Notes for April 5, 2016 Dynamic programming, continued (HMMs); Iterative improvement Bernard Moret Dynamic Programming (continued) Finite Markov models A finite Markov model is a discrete-time (event-driven) stochastic model defined on a set of states S = {s 1,s 2,...,s n } through a transition probability matrix T = (t i j ), where t i j is the probability of moving from state s i to state s j. (In practice, many of the entries in T are set to zero: only certain transitions are considered in the model.) T is a (right) stochastic matrix, i.e., each t i j is a real number in [0,1] and every row sums to 1. A finite Markov model is sometimes also called a stochastic finite automaton. As with finite automata, it is convenient to add an output function to the model. While we could associate an output with a transition, the normal usage is to associate an output with a state the equivalent of a Mealy machine for finite automata. Suppose the output is a character from the alphabet Σ, then we add a second (right) stochastic matrix E =(e i j ), the emission matrix, where e i j is the probability of producing (emitting) the jth character in Σ when in state s i. If we run the model, it will produce some output string; for convenience, we assume that this string is finite. Notice that we are looking at a very large number of parameters: in the worst case, the n n transition matrix has n(n 1) parameters, to which we must add n ( Σ 1) parameters for the emission matrix. In practice, therefore, we assume that we know the finite automaton (i.e., we know the number of states and we know which transitions are assumed to have zero probability), but not its parameters: we will attempt to infer (learn) the parameters from collected data. So what can we collect about the model? Obviously, we can collect data representing the output; so denote by x the observed output string, by x i the character in the ith position of x, and by m the length of x. (The model may be capable of producing strings of unbounded length, but of course we can make only bounded observations.) In a limited number of problems, we may also be able to observe directly the state of the model, thereby giving rise, for the same m, to a string of observed states y=y 1 Y 2...y m. Hidden Markov Models For most problems arising from nature, the actual state of the process under observation is not observable indeed, it may not be well defined, since the model is certainly a simplification of nature. However, since we are observing the process, we can certainly collect the output. In our terminology, we can get x, but not y; when such is the case, we call the model a Hidden Markov Model HMM for short. 1
2 HMMs are found everyhwere, but are perhaps most common in two areas. They completely changed the area of voice recognition, an area that that been lavishly funded, mostly in vain, for 30 years by the military. Now we have Alexa, Cortana, Siri, and other personal assistants, and the voice recognition part is completely routine the challenge is in understanding the nature of the requests and formulating appropriate responses. The other area is computational biology, where HMMs are used to analyze genomes (at several billion elements per, these are quite long strings), on much the same basis: let the HMM do the work of recognizing patterns rather than attempt develop specialized algorithms that keep failing to generalize properly. One can use an HMM as a recognizer: by formulating and training several HMMs, one for each class of outcomes, one can use these HMMs to compute the probability that each would be generated some new sequence of observations and then assign that new sequence to the class most likely to have generated it. One can also study the innards of the machine and ask about the probability of occupation of a certain state, the most likely path through the HMM for a given sequence of observations, etc. These all require knowledge of the entire model, including the values of all of its numerous parameters, something that is not usually preset by the designer, but inferred (through training) from actual data. Hence another fundamental question about an HMM is how to infer the values t i j and e ik from observed data. This last is done today through the Baum-Welch algorithm, a special case of EM (expectation-maximization), in effect an approximation algorithm based on iterative improvement. We shall return to it later in the course. The former (classification and detailed statistics on the functioning), however, is done through dynamic programming and so we tackle them now. We formalize three related questions: 1. How do we reconstruct (and compute the likelihood of) the most likely state sequence for the given HMM and given output sequence? 2. How do we compute the likelihood of a given sequence? 3. How do we compute the likelihood of a particular state at a particular step for the observed output sequence? All three of these questions can be answered by the same basic dynamic program, with small variations. We will build a table with one row for each state and one column for each position in the observed output sequence thus our table will be n m. (Note that we may want to introduce two additional states, i.e., two additional rows, and one additional column; the extra states are a start state and an end state, the extra column corresponds to position 0 before the position of the first characters in the output string. These additions do not alter the following, but prevent having to treat special rows or columns differently from others and thus are likely to improve efficiency.) 1. The solution here is the so-called Viterbi algorithm. We will fill the table column by column, starting at column 0 and ending at column m; entry i j in the table, that is, 2
3 P v (i, j), will denote the probability of the most likely path to state s i after reading the jth character in x. We can compute P(i, j) according to the following recurrence: ( P v (i, j)=max ali e i,x j P v (l, j 1) ) l In words, we just look one step back along the path, to a previous entry P v (l, j 1) for some previous state l, and compute the probability of the extension from that previous step to the current state s i at step j with symbol x j emitted, retaining the largest one. It should be noted, that, in spite of appearances, what is computed is really a joint probability, not (at least not intentionally) a conditional probability. That is, the recurrence computes the maximum value of the probability p(π,x), where x is the output string and π is a path through the states of the HMM that produces that string. However, it is not hard to see that computing the max (over choices of π) of p(π,x) is equivalent to computing the max of p(π x), the conditional (posterior) probability. 2. To get the likelihood of the output sequence, we could look at every state path that can generate it, compute the probability of each such path, then sum them all to obtain p(x) = π p(π,x), using the notation from (i); but this takes time proportional to the number of paths and thus could take exponential time. Instead, we implicitly look at all paths every time we add another output character. The recurrence is nearly identical to the Viterbi recurrence we simply replace the max operator by an addition. ( P f (i, j)= ali e i,x j P f (l, j 1) ) l (Note that the emission probability is independent of the summation index and so can be factored.) The dynamic program resulting from this recurrence is often called the forward algorithm, hence my choice of P f for this function. 3. Here we really want to compute p(s i, j x), the probability that, given that the HMM produced the output sequence x, it was in state s i at step j. Because this is a conditional probability (unlike the two computed above), we proceed somewhat indirectly, using joint probabilities that we do know how to compute. We have p(s i, j,x)= p(s i,x 1 x 2 x j ) p(x j+1 x j+2 x m s i,x 1 x 2 x j ) Note that the first term in the product is exactly P f (i, j) in the forward algorithm, so we know how to compute it. The second term can be simplified: thanks to the Markov (memoryless) property, the dependency on x 1 x 2 x j does not exist, so the second term now reduces to p(x j+1 x j+2 x m s i ), and we can easily compute that term by the same dynamic program again, but this time running it backward (which gives us the condition for free). The recurrence is then P b (i, j)= a li e i,x j+1 P b (l, j+1) l 3
4 where the b subscript denotes that the backward version. Now we simply have p(s i, j x)= P f(i, j) P b (i, j) P(x) What is the running time of each of these three algorithms? All have to fill in the entire n m table; moreover, in order to fill it, all have to look at all transitions into (or out of) each state. Thus, to process one column of the table, all transitions of the HMM must be examined and used in the recurrence, so that the cost of processing one entire column is proportional to α n, the number of nonzero entries in the transition matrix. Thus the overall running time of the algorithms is Θ(α n m); this can be as low as order mn if α (which is twice the average degree of a node in the state transition diagram) is a constant and as high as order mn 2 if α is some fraction of n (as in a dense state graph). Clearly, there are other questions we might ask about the model itself, such as the probability that the HMM is in some given state at any time during the process, the probability that a particular transition is used at any time during the process, and many others. Most, if not all, such questions can be answered through dynamic programming approaches similar to the three above. Iterative Improvement Stable Marriage Imagine you run a match-making service and introduce 20 young men and 20 young women with the goal of eventually forming 20 happy couples. After some period of getting acquainted, you eventually ask each woman to rank order all 20 men, from first choice to last choice as a potential husband, and similarly you ask each man to rank order all 20 women, from first choice to last choice as a potential wife. A marriage of woman y to man x is said to be stable if there does not exist any other couple (x,y ) such that either woman y prefers x to her own husband and man x prefers y to his own wife, or (the symmetric situation) men x prefers y to his own wife and woman y prefers x to her own husband. (If either of these two situations occurs, then obviously there an incentive for a swap of spouses, which would remedy the situation hence the marriages (x,y) and (x,y ) would not be stable.) Phrased as a marital endeavor, the problem seems silly, but in fact it has quite a range of applications, although in most applications the lists of preferences will not include all members of the other subset. The best known application in the US is The Match for medical residencies: medical students about to graduate visit several residency programs (teaching hospitals with openings in this or that area of medicine) all over the USA, while these programs receive visits from a (much larger) number of medical students and interview these students. (Most residency programs have more than one position in each medical area, hence the larger number of students.) At the end of this period of courtship, each medical student files with a central authority ( a list of her/his preferences (i.e., a rank ordering of the programs visited), while each residency program files with the same central authority a rank-ordered 4
5 list of all students who visited. Then. on a fatal day known simply as MATCH DAY. a computer program computes a stable matching, that is a list of pairs (student, residency position), and both students and residency programs are notified of the outcome. Nearly students participate in The Match each year and almost all residency positions are filled through it. Gale and Shapley, the scientists who formulated the stable marriage problem (well, one of the two, since the other had died) and the scientist who applied it to the medical residency problem (Roth) received the Nobel prize in economics in 2012 for their work on this problem. The work dates all the way back to 1962, when Gale and Shapley published a famous paper entitled College admissions and the stability of marriage (American Mathematics Monthly 69:9 15), in which they defined a notion of stable marriage (matching/pairing/assignment/etc.) and showed that, in matching seekers to offerers, one could produced stable matchings that were optimal for seekers or stable matchings that were optimal for offerers. (In the case of the medical match, the stable matching produced is optimal for medical graduates.) The formulation they gave was for college admissions, but they also introduced a formulation in the context of marriage. The algorithm used in the match proceeds in rounds as follows: In the first round, every man proposes to the top woman on his list. Each woman accepts the best proposal she receives. (Naturally, a number of women will receive multiple proposals while some will receive none.) In the second round, every man not yet engaged proposes to the second choice on his list, regardless of whether she is already engaged or not. Each woman retains the best proposal she has received so far; this may entail breaking the engagement she made at the previous round because she received a proposal from a better-ranked man in this round, causing a man who was engaged at the end of the first round to be unengaged at the end of the second round. In each subsequent round, each unengaged man proposes to the next woman (engaged or not) on his list; each woman retains the best proposal she has received across all rounds so far. This continues until every man (and thus also every woman) is engaged, at which point all engagements become marriages. Readers will recognize this algorithm as the prevailing social custom in Western civilization for many centuries (up to the 19th century and even later): women could not make a move of their own (although they could certainly drop hints!), but had to wait on the pleasure of men to declare themselves. Unsurprisingly, like much about human civilizations, this proposal algorithm is optimal for men; it is also pessimal for women. (Simply flipping the roles will give us an algorithm optimal for women and pessimal for men; it turns out, however, that designing an algorithm that is fair to both sexes is much harder: it was not until the 21st century that the first such algorithm was published.) The match has one major difference from the marriage version: the rankings of the medical graduates include only the 5 15 residency programs where they interviewed and the rankings of the residency programs, while larger, also fall far short of listing all medical graduates. Fortunately, this difference is easily handled through the same algorithm. 5
6 That this algorithm terminates is clear: at round i, every unengaged man proposes to the ith woman on his list, so the process stops after at most n rounds. That it ends with n marriages is also clear, and due to the fact that everyone is willing to marry anyone of the opposite sex there are preferences, but there cannot be rejections. That the marriages thus chosen are stable is due to the process: if x ends up married to y, but prefers y, then he must have proposed to y before proposing to y and have been rejected, which means that y was already engaged at that time to someone higher on her list than x. Since the ranking of a woman s partner can only go up from the first engagement, y is now married to someone at least as high on her list as the man she already preferred to x during the rounds of proposals and so is not interested in switching. A similar line of reasoning holds for a man that y might prefer to x. The algorithm may result in a quadratic number of proposals (n rounds, with only one additional engagement per round). 6
Ma/CS 6b Class 3: Stable Matchings
Ma/CS 6b Class 3: Stable Matchings α p 5 p 12 p 15 q 1 q 7 q 12 β By Adam Sheffer Neighbor Sets Let G = V 1 V 2, E be a bipartite graph. For any vertex a V 1, we define the neighbor set of a as N a = u
More informationMa/CS 6b Class 3: Stable Matchings
Ma/CS 6b Class 3: Stable Matchings α p 5 p 12 p 15 q 1 q 7 q 12 By Adam Sheffer Reminder: Alternating Paths Let G = V 1 V 2, E be a bipartite graph, and let M be a matching of G. A path is alternating
More informationComputational Genomics and Molecular Biology, Fall
Computational Genomics and Molecular Biology, Fall 2011 1 HMM Lecture Notes Dannie Durand and Rose Hoberman October 11th 1 Hidden Markov Models In the last few lectures, we have focussed on three problems
More information1. REPRESENTATIVE PROBLEMS
1. REPRESENTATIVE PROBLEMS stable matching five representative problems Lecture slides by Kevin Wayne Copyright 2005 Pearson-Addison Wesley Copyright 2013 Kevin Wayne http://www.cs.princeton.edu/~wayne/kleinberg-tardos
More informationBipartite Matchings and Stable Marriage
Bipartite Matchings and Stable Marriage Meghana Nasre Department of Computer Science and Engineering Indian Institute of Technology, Madras Faculty Development Program SSN College of Engineering, Chennai
More informationSubramanian s stable matching algorithm
Subramanian s stable matching algorithm Yuval Filmus November 2011 Abstract Gale and Shapley introduced the well-known stable matching problem in 1962, giving an algorithm for the problem. Subramanian
More informationMatching Theory and the Allocation of Kidney Transplantations
University of Utrecht Bachelor Thesis Matching Theory and the Allocation of Kidney Transplantations Kim de Bakker Supervised by Dr. M. Ruijgrok 14 June 2016 Introduction Matching Theory has been around
More informationStable matching. Carlos Hurtado. July 5th, Department of Economics University of Illinois at Urbana-Champaign
Stable matching Carlos Hurtado Department of Economics University of Illinois at Urbana-Champaign hrtdmrt2@illinois.edu July 5th, 2017 C. Hurtado (UIUC - Economics) Game Theory On the Agenda 1 Introduction
More informationWhat do you do when you can t use money to solve your problems?
Markets without money What do you do when you can t use money to solve your problems? Matching: heterosexual men and women marrying in a small town, students matching to universities, workers to jobs where
More informationHidden Markov Models. Ivan Gesteira Costa Filho IZKF Research Group Bioinformatics RWTH Aachen Adapted from:
Hidden Markov Models Ivan Gesteira Costa Filho IZKF Research Group Bioinformatics RWTH Aachen Adapted from: www.ioalgorithms.info Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm
More information1 Definitions and Things You Know
We will discuss an algorithm for finding stable matchings (not the one you re probably familiar with). The Instability Chaining Algorithm is the most similar algorithm in the literature to the one actually
More informationCSC2556. Lecture 5. Matching - Stable Matching - Kidney Exchange [Slides : Ariel D. Procaccia]
CSC2556 Lecture 5 Matching - Stable Matching - Kidney Exchange [Slides : Ariel D. Procaccia] CSC2556 - Nisarg Shah 1 Announcements The assignment is up! It is complete, and no more questions will be added.
More informationCISC 889 Bioinformatics (Spring 2004) Hidden Markov Models (II)
CISC 889 Bioinformatics (Spring 24) Hidden Markov Models (II) a. Likelihood: forward algorithm b. Decoding: Viterbi algorithm c. Model building: Baum-Welch algorithm Viterbi training Hidden Markov models
More information1. REPRESENTATIVE PROBLEMS
1. REPRESENTATIVE PROBLEMS stable matching five representative problems Special thanks to Kevin Wayne for sharing the slides Copyright 2005 Pearson-Addison Wesley Last updated on 15/9/12 下午 10:33 1. REPRESENTATIVE
More informationMatching. Terence Johnson. April 17, University of Notre Dame. Terence Johnson (ND) Matching April 17, / 41
Matching Terence Johnson University of Notre Dame April 17, 2018 Terence Johnson (ND) Matching April 17, 2018 1 / 41 Markets without money What do you do when you can t use money to solve your problems?
More informationMatchings in Graphs. Definition 3 A matching N in G is said to be stable if it does not contain a blocking pair.
Matchings in Graphs Lecturer: Scribe: Prajakta Jose Mathew Meeting: 6 11th February 2010 We will be considering finite bipartite graphs. Think of one part of the vertex partition as representing men M,
More information11.3 Decoding Algorithm
11.3 Decoding Algorithm 393 For convenience, we have introduced π 0 and π n+1 as the fictitious initial and terminal states begin and end. This model defines the probability P(x π) for a given sequence
More informationMultiplex network inference
(using hidden Markov models) University of Cambridge Bioinformatics Group Meeting 11 February 2016 Words of warning Disclaimer These slides have been produced by combining & translating two of my previous
More informationLecture 9. Intro to Hidden Markov Models (finish up)
Lecture 9 Intro to Hidden Markov Models (finish up) Review Structure Number of states Q 1.. Q N M output symbols Parameters: Transition probability matrix a ij Emission probabilities b i (a), which is
More information8: Hidden Markov Models
8: Hidden Markov Models Machine Learning and Real-world Data Simone Teufel and Ann Copestake Computer Laboratory University of Cambridge Lent 2017 Last session: catchup 1 Research ideas from sentiment
More informationAn Introduction to Bioinformatics Algorithms Hidden Markov Models
Hidden Markov Models Hidden Markov Models Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm Forward-Backward Algorithm Profile HMMs HMM Parameter Estimation Viterbi training
More informationCS320 Algorithms: Theory and Practice. Course Introduction
Course Objectives CS320 Algorithms: Theory and Practice Algorithms: Design strategies for algorithmic problem solving Course Introduction "For me, great algorithms are the poetry of computation. Just like
More informationTwo-Sided Matching. Terence Johnson. December 1, University of Notre Dame. Terence Johnson (ND) Two-Sided Matching December 1, / 47
Two-Sided Matching Terence Johnson University of Notre Dame December 1, 2017 Terence Johnson (ND) Two-Sided Matching December 1, 2017 1 / 47 Markets without money What do you do when you can t use money
More informationHidden Markov models
Hidden Markov models Charles Elkan November 26, 2012 Important: These lecture notes are based on notes written by Lawrence Saul. Also, these typeset notes lack illustrations. See the classroom lectures
More informationHidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010
Hidden Markov Models Aarti Singh Slides courtesy: Eric Xing Machine Learning 10-701/15-781 Nov 8, 2010 i.i.d to sequential data So far we assumed independent, identically distributed data Sequential data
More informationAnalysis of Algorithms Fall Some Representative Problems Stable Matching
Analysis of Algorithms Fall 2017 Some Representative Problems Stable Matching Mohammad Ashiqur Rahman Department of Computer Science College of Engineering Tennessee Tech University Matching Med-school
More informationCS 6901 (Applied Algorithms) Lecture 2
CS 6901 (Applied Algorithms) Lecture 2 Antonina Kolokolova September 15, 2016 1 Stable Matching Recall the Stable Matching problem from the last class: there are two groups of equal size (e.g. men and
More informationNatural Language Processing Prof. Pushpak Bhattacharyya Department of Computer Science & Engineering, Indian Institute of Technology, Bombay
Natural Language Processing Prof. Pushpak Bhattacharyya Department of Computer Science & Engineering, Indian Institute of Technology, Bombay Lecture - 21 HMM, Forward and Backward Algorithms, Baum Welch
More informationPRIMES Math Problem Set
PRIMES Math Problem Set PRIMES 017 Due December 1, 01 Dear PRIMES applicant: This is the PRIMES 017 Math Problem Set. Please send us your solutions as part of your PRIMES application by December 1, 01.
More informationHidden Markov Models. By Parisa Abedi. Slides courtesy: Eric Xing
Hidden Markov Models By Parisa Abedi Slides courtesy: Eric Xing i.i.d to sequential data So far we assumed independent, identically distributed data Sequential (non i.i.d.) data Time-series data E.g. Speech
More informationSTA 414/2104: Machine Learning
STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Hidden Markov Models Barnabás Póczos & Aarti Singh Slides courtesy: Eric Xing i.i.d to sequential data So far we assumed independent, identically distributed
More informationorder is number of previous outputs
Markov Models Lecture : Markov and Hidden Markov Models PSfrag Use past replacements as state. Next output depends on previous output(s): y t = f[y t, y t,...] order is number of previous outputs y t y
More informationAlgorithms. [Knuth, TAOCP] An algorithm is a finite, definite, effective procedure, with some input and some output.
Algorithms Algorithm. [webster.com] A procedure for solving a mathematical problem (as of finding the greatest common divisor) in a finite number of steps that frequently involves repetition of an operation.
More informationWe set up the basic model of two-sided, one-to-one matching
Econ 805 Advanced Micro Theory I Dan Quint Fall 2009 Lecture 18 To recap Tuesday: We set up the basic model of two-sided, one-to-one matching Two finite populations, call them Men and Women, who want to
More informationHidden Markov Models. Three classic HMM problems
An Introduction to Bioinformatics Algorithms www.bioalgorithms.info Hidden Markov Models Slides revised and adapted to Computational Biology IST 2015/2016 Ana Teresa Freitas Three classic HMM problems
More informationO 3 O 4 O 5. q 3. q 4. Transition
Hidden Markov Models Hidden Markov models (HMM) were developed in the early part of the 1970 s and at that time mostly applied in the area of computerized speech recognition. They are first described in
More informationStatistical Methods for NLP
Statistical Methods for NLP Sequence Models Joakim Nivre Uppsala University Department of Linguistics and Philology joakim.nivre@lingfil.uu.se Statistical Methods for NLP 1(21) Introduction Structured
More informationMACHINE LEARNING 2 UGM,HMMS Lecture 7
LOREM I P S U M Royal Institute of Technology MACHINE LEARNING 2 UGM,HMMS Lecture 7 THIS LECTURE DGM semantics UGM De-noising HMMs Applications (interesting probabilities) DP for generation probability
More informationCSCE 471/871 Lecture 3: Markov Chains and
and and 1 / 26 sscott@cse.unl.edu 2 / 26 Outline and chains models (s) Formal definition Finding most probable state path (Viterbi algorithm) Forward and backward algorithms State sequence known State
More informationStatistical NLP: Hidden Markov Models. Updated 12/15
Statistical NLP: Hidden Markov Models Updated 12/15 Markov Models Markov models are statistical tools that are useful for NLP because they can be used for part-of-speech-tagging applications Their first
More informationHIDDEN MARKOV MODELS
HIDDEN MARKOV MODELS Outline CG-islands The Fair Bet Casino Hidden Markov Model Decoding Algorithm Forward-Backward Algorithm Profile HMMs HMM Parameter Estimation Viterbi training Baum-Welch algorithm
More informationMath 301: Matchings in Graphs
Math 301: Matchings in Graphs Mary Radcliffe 1 Definitions and Basics We begin by first recalling some basic definitions about matchings. A matching in a graph G is a set M = {e 1, e 2,..., e k } of edges
More informationHidden Markov Models
Hidden Markov Models CI/CI(CS) UE, SS 2015 Christian Knoll Signal Processing and Speech Communication Laboratory Graz University of Technology June 23, 2015 CI/CI(CS) SS 2015 June 23, 2015 Slide 1/26 Content
More informationHidden Markov Models. Hosein Mohimani GHC7717
Hidden Markov Models Hosein Mohimani GHC7717 hoseinm@andrew.cmu.edu Fair et Casino Problem Dealer flips a coin and player bets on outcome Dealer use either a fair coin (head and tail equally likely) or
More informationHidden Markov Models
Hidden Markov Models Slides revised and adapted to Bioinformática 55 Engª Biomédica/IST 2005 Ana Teresa Freitas Forward Algorithm For Markov chains we calculate the probability of a sequence, P(x) How
More informationAn Introduction to Bioinformatics Algorithms Hidden Markov Models
Hidden Markov Models Outline 1. CG-Islands 2. The Fair Bet Casino 3. Hidden Markov Model 4. Decoding Algorithm 5. Forward-Backward Algorithm 6. Profile HMMs 7. HMM Parameter Estimation 8. Viterbi Training
More informationCheating to Get Better Roommates in a Random Stable Matching
Cheating to Get Better Roommates in a Random Stable Matching Chien-Chung Huang Technical Report 2006-582 Dartmouth College Sudikoff Lab 6211 for Computer Science Hanover, NH 03755, USA villars@cs.dartmouth.edu
More informationGame Theory: Lecture #5
Game Theory: Lecture #5 Outline: Stable Matchings The Gale-Shapley Algorithm Optimality Uniqueness Stable Matchings Example: The Roommate Problem Potential Roommates: {A, B, C, D} Goal: Divide into two
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More informationMatching Theory. Mihai Manea. Based on slides by Fuhito Kojima. MIT
Matching Theory Mihai Manea MIT Based on slides by Fuhito Kojima. Market Design Traditional economics focuses mostly on decentralized markets. Recently, economists are helping to design economic institutions
More informationGreedy Homework Problems
CS 1510 Greedy Homework Problems 1. Consider the following problem: INPUT: A set S = {(x i, y i ) 1 i n} of intervals over the real line. OUTPUT: A maximum cardinality subset S of S such that no pair of
More informationHidden Markov Models Part 2: Algorithms
Hidden Markov Models Part 2: Algorithms CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Hidden Markov Model An HMM consists of:
More informationHidden Markov Models
Hidden Markov Models Outline 1. CG-Islands 2. The Fair Bet Casino 3. Hidden Markov Model 4. Decoding Algorithm 5. Forward-Backward Algorithm 6. Profile HMMs 7. HMM Parameter Estimation 8. Viterbi Training
More informationAlgorithms: COMP3121/3821/9101/9801
Algorithms: COMP311/381/9101/9801 Aleks Ignjatović, ignjat@cse.unsw.edu.au office: 504 (CSE building); phone: 5-6659 Course Admin: Amin Malekpour, a.malekpour@unsw.edu.au School of Computer Science and
More informationWeights in stable marriage problems increase manipulation opportunities
Weights in stable marriage problems increase manipulation opportunities Maria Silvia Pini 1, Francesca Rossi 1, Kristen Brent Venable 1, Toby Walsh 2 1 : Department of Pure and Applied Mathematics, University
More informationAn improved approximation algorithm for the stable marriage problem with one-sided ties
Noname manuscript No. (will be inserted by the editor) An improved approximation algorithm for the stable marriage problem with one-sided ties Chien-Chung Huang Telikepalli Kavitha Received: date / Accepted:
More informationDiscrete Mathematics and Probability Theory Fall 2014 Anant Sahai Midterm 1
EECS 70 Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Midterm 1 Exam location: 10 Evans, Last name starting with A-B or R-T PRINT your student ID: PRINT AND SIGN your name:, (last)
More informationHidden Markov Models
Hidden Markov Models Slides mostly from Mitch Marcus and Eric Fosler (with lots of modifications). Have you seen HMMs? Have you seen Kalman filters? Have you seen dynamic programming? HMMs are dynamic
More information8: Hidden Markov Models
8: Hidden Markov Models Machine Learning and Real-world Data Helen Yannakoudakis 1 Computer Laboratory University of Cambridge Lent 2018 1 Based on slides created by Simone Teufel So far we ve looked at
More informationMarkov Chains and Hidden Markov Models. = stochastic, generative models
Markov Chains and Hidden Markov Models = stochastic, generative models (Drawing heavily from Durbin et al., Biological Sequence Analysis) BCH339N Systems Biology / Bioinformatics Spring 2016 Edward Marcotte,
More informationThe key is that there are two disjoint populations, and everyone in the market is on either one side or the other
Econ 805 Advanced Micro Theory I Dan Quint Fall 2009 Lecture 17 So... two-sided matching markets. First off, sources. I ve updated the syllabus for the next few lectures. As always, most of the papers
More informationHidden Markov Model. Ying Wu. Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208
Hidden Markov Model Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1/19 Outline Example: Hidden Coin Tossing Hidden
More informationVL Algorithmen und Datenstrukturen für Bioinformatik ( ) WS15/2016 Woche 16
VL Algorithmen und Datenstrukturen für Bioinformatik (19400001) WS15/2016 Woche 16 Tim Conrad AG Medical Bioinformatics Institut für Mathematik & Informatik, Freie Universität Berlin Based on slides by
More informationSTABLE MARRIAGE PROBLEM WITH TIES AND INCOMPLETE BOUNDED LENGTH PREFERENCE LIST UNDER SOCIAL STABILITY
STABLE MARRIAGE PROBLEM WITH TIES AND INCOMPLETE BOUNDED LENGTH PREFERENCE LIST UNDER SOCIAL STABILITY Ashish Shrivastava and C. Pandu Rangan Department of Computer Science and Engineering, Indian Institute
More informationPlan for today. ! Part 1: (Hidden) Markov models. ! Part 2: String matching and read mapping
Plan for today! Part 1: (Hidden) Markov models! Part 2: String matching and read mapping! 2.1 Exact algorithms! 2.2 Heuristic methods for approximate search (Hidden) Markov models Why consider probabilistics
More informationDynamic Approaches: The Hidden Markov Model
Dynamic Approaches: The Hidden Markov Model Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Machine Learning: Neural Networks and Advanced Models (AA2) Inference as Message
More informationSTATS 306B: Unsupervised Learning Spring Lecture 5 April 14
STATS 306B: Unsupervised Learning Spring 2014 Lecture 5 April 14 Lecturer: Lester Mackey Scribe: Brian Do and Robin Jia 5.1 Discrete Hidden Markov Models 5.1.1 Recap In the last lecture, we introduced
More informationHidden Markov Models
CS769 Spring 2010 Advanced Natural Language Processing Hidden Markov Models Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu 1 Part-of-Speech Tagging The goal of Part-of-Speech (POS) tagging is to label each
More information1 What is a hidden Markov model?
1 What is a hidden Markov model? Consider a Markov chain {X k }, where k is a non-negative integer. Suppose {X k } embedded in signals corrupted by some noise. Indeed, {X k } is hidden due to noise and
More informationGenome 373: Hidden Markov Models II. Doug Fowler
Genome 373: Hidden Markov Models II Doug Fowler Review From Hidden Markov Models I What does a Markov model describe? Review From Hidden Markov Models I A T A Markov model describes a random process of
More informationHidden Markov Models. x 1 x 2 x 3 x N
Hidden Markov Models 1 1 1 1 K K K K x 1 x x 3 x N Example: The dishonest casino A casino has two dice: Fair die P(1) = P() = P(3) = P(4) = P(5) = P(6) = 1/6 Loaded die P(1) = P() = P(3) = P(4) = P(5)
More informationIntroduction to Hidden Markov Models for Gene Prediction ECE-S690
Introduction to Hidden Markov Models for Gene Prediction ECE-S690 Outline Markov Models The Hidden Part How can we use this for gene prediction? Learning Models Want to recognize patterns (e.g. sequence
More informationHMM: Parameter Estimation
I529: Machine Learning in Bioinformatics (Spring 2017) HMM: Parameter Estimation Yuzhen Ye School of Informatics and Computing Indiana University, Bloomington Spring 2017 Content Review HMM: three problems
More informationHidden Markov Models
Hidden Markov Models Lecture Notes Speech Communication 2, SS 2004 Erhard Rank/Franz Pernkopf Signal Processing and Speech Communication Laboratory Graz University of Technology Inffeldgasse 16c, A-8010
More informationHidden Markov Models The three basic HMM problems (note: change in notation) Mitch Marcus CSE 391
Hidden Markov Models The three basic HMM problems (note: change in notation) Mitch Marcus CSE 391 Parameters of an HMM States: A set of states S=s 1, s n Transition probabilities: A= a 1,1, a 1,2,, a n,n
More informationBasic math for biology
Basic math for biology Lei Li Florida State University, Feb 6, 2002 The EM algorithm: setup Parametric models: {P θ }. Data: full data (Y, X); partial data Y. Missing data: X. Likelihood and maximum likelihood
More informationStable Marriage with Ties and Bounded Length Preference Lists
Stable Marriage with Ties and Bounded Length Preference Lists Robert W. Irving, David F. Manlove, and Gregg O Malley Department of Computing Science, University of Glasgow, Glasgow G12 8QQ, UK. Email:
More informationStrategic Behavior and Manipulation in Gender- Neutral Matching Algorithms
Rochester Institute of Technology RIT Scholar Works Theses Thesis/Dissertation Collections 12-2018 Strategic Behavior and Manipulation in Gender- Neutral Matching Algorithms Sanjay Varma Rudraraju sr2567@rit.edu
More informationLecture 11: Hidden Markov Models
Lecture 11: Hidden Markov Models Cognitive Systems - Machine Learning Cognitive Systems, Applied Computer Science, Bamberg University slides by Dr. Philip Jackson Centre for Vision, Speech & Signal Processing
More informationMachine Learning for natural language processing
Machine Learning for natural language processing Hidden Markov Models Laura Kallmeyer Heinrich-Heine-Universität Düsseldorf Summer 2016 1 / 33 Introduction So far, we have classified texts/observations
More informationCPSC 320 Sample Solution, Reductions and Resident Matching: A Residentectomy
CPSC 320 Sample Solution, Reductions and Resident Matching: A Residentectomy August 25, 2017 A group of residents each needs a residency in some hospital. A group of hospitals each need some number (one
More informationHidden Markov Models. x 1 x 2 x 3 x K
Hidden Markov Models 1 1 1 1 2 2 2 2 K K K K x 1 x 2 x 3 x K Viterbi, Forward, Backward VITERBI FORWARD BACKWARD Initialization: V 0 (0) = 1 V k (0) = 0, for all k > 0 Initialization: f 0 (0) = 1 f k (0)
More informationMath 350: An exploration of HMMs through doodles.
Math 350: An exploration of HMMs through doodles. Joshua Little (407673) 19 December 2012 1 Background 1.1 Hidden Markov models. Markov chains (MCs) work well for modelling discrete-time processes, or
More informationDiscrete Mathematics and Probability Theory Fall 2014 Anant Sahai Homework 3. This homework is due September 22, 2014, at 12:00 noon.
EECS 70 Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Homework 3 This homework is due September 22, 2014, at 12:00 noon. 1. Propose-and-Reject Lab In this week s Virtual Lab, we will
More informationStephen Scott.
1 / 27 sscott@cse.unl.edu 2 / 27 Useful for modeling/making predictions on sequential data E.g., biological sequences, text, series of sounds/spoken words Will return to graphical models that are generative
More information1 Probabilities. 1.1 Basics 1 PROBABILITIES
1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability
More informationWe Live in Exciting Times. CSCI-567: Machine Learning (Spring 2019) Outline. Outline. ACM (an international computing research society) has named
We Live in Exciting Times ACM (an international computing research society) has named CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Apr. 2, 2019 Yoshua Bengio,
More informationComputational Genomics and Molecular Biology, Fall
Computational Genomics and Molecular Biology, Fall 2014 1 HMM Lecture Notes Dannie Durand and Rose Hoberman November 6th Introduction In the last few lectures, we have focused on three problems related
More informationHidden Markov Models. Terminology, Representation and Basic Problems
Hidden Markov Models Terminology, Representation and Basic Problems Data analysis? Machine learning? In bioinformatics, we analyze a lot of (sequential) data (biological sequences) to learn unknown parameters
More informationHidden Markov Models for biological sequence analysis
Hidden Markov Models for biological sequence analysis Master in Bioinformatics UPF 2017-2018 http://comprna.upf.edu/courses/master_agb/ Eduardo Eyras Computational Genomics Pompeu Fabra University - ICREA
More informationAlgorithm Design and Analysis
Algorithm Design and Analysis LECTURE 2 Analysis of Stable Matching Asymptotic Notation Adam Smith Stable Matching Problem Goal: Given n men and n women, find a "suitable" matching. Participants rate members
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression
More informationAlgorithm Design and Analysis
Algorithm Design and Analysis LECTURE 2 Analysis of Algorithms Stable matching problem Asymptotic growth Adam Smith Stable Matching Problem Unstable pair: man m and woman w are unstable if m prefers w
More informationSEQUENTIAL ENTRY IN ONE-TO-ONE MATCHING MARKETS
REVISTA DE LA UNIÓN MATEMÁTICA ARGENTINA Vol. 54, No. 2, 2013, Pages 1 14 Published online: December 21, 2013 SEQUENTIAL ENTRY IN ONE-TO-ONE MATCHING MARKETS BEATRIZ MILLÁN Abstract. We study in one-to-one
More informationLecture 12: Algorithms for HMMs
Lecture 12: Algorithms for HMMs Nathan Schneider (some slides from Sharon Goldwater; thanks to Jonathan May for bug fixes) ENLP 26 February 2018 Recap: tagging POS tagging is a sequence labelling task.
More information1. STABLE MATCHING. stable matching problem Gale Shapley algorithm hospital optimality context
1. STABLE MATCHING stable matching problem Gale Shapley algorithm hospital optimality context Lecture slides by Kevin Wayne Copyright 2005 Pearson-Addison Wesley http://www.cs.princeton.edu/~wayne/kleinberg-tardos
More informationRecall: Modeling Time Series. CSE 586, Spring 2015 Computer Vision II. Hidden Markov Model and Kalman Filter. Modeling Time Series
Recall: Modeling Time Series CSE 586, Spring 2015 Computer Vision II Hidden Markov Model and Kalman Filter State-Space Model: You have a Markov chain of latent (unobserved) states Each state generates
More informationMarkov Chains and Hidden Markov Models. COMP 571 Luay Nakhleh, Rice University
Markov Chains and Hidden Markov Models COMP 571 Luay Nakhleh, Rice University Markov Chains and Hidden Markov Models Modeling the statistical properties of biological sequences and distinguishing regions
More informationHidden Markov Models for biological sequence analysis I
Hidden Markov Models for biological sequence analysis I Master in Bioinformatics UPF 2014-2015 Eduardo Eyras Computational Genomics Pompeu Fabra University - ICREA Barcelona, Spain Example: CpG Islands
More information