Sequential Supervised Learning
|
|
- Bethany Barber
- 5 years ago
- Views:
Transcription
1 Sequential Supervised Learning
2 Many Application Problems Require Sequential Learning Part-of of-speech Tagging Information Extraction from the Web Text-to to-speech Mapping
3 Part-of of-speech Tagging Given an English sentence, can we assign a part of speech to each word? Do you want fries with that? <verb pron verb noun prep pron>
4 Information Extraction from the Web <dl><dt><b>srinivasan Seshan</b> (Carnegie Mellon University) <dt><a href= ><i>making Virtual Worlds Real</i></a><dt>Tuesday, June, 00<dd>:00 PM, Sieg<dd>Research Seminar * * * name name * * affiliation affiliation affiliation * * * * title title title title * * * date date date date * time time * location location * event-type event-type
5 Text-to to-speech Mapping photograph =>
6 Sequential Supervised Learning (SSL) Given: A set of training examples of the form (X( i,y i ), where X i = hx i,,, x i,t ii and Y i = hy i,,, y i,t ii are sequences of length T i Find: A function f for predicting new sequences: Y = f(x).
7 Examples of Sequential Supervised Learning Domain Part-of of-speech Tagging Information Extraction Test-to to-speech Mapping Input X i sequence of words sequence of tokens sequence of letters Output Y i sequence of parts of speech sequence of field labels {name, } sequence phonemes
8 Two Kinds of Relationships y y y x x x Vertical relationship between the x t s and y t s Example: Friday is usually a date Horizontal relationships among the y t s Example: name is usually followed by affiliation SSL can (and should) exploit both kinds of information
9 Existing Methods Hacks Sliding windows Recurrent sliding windows Hidden Markov models joint distribution: P(X,Y) Conditional Random Fields conditional distribution: P(Y X) Discriminant Methods: HM-SVMs, MMMs, voted perceptrons discriminant function: f(y; X)
10 Sliding Windows Sliding Windows that that with with fries fries want want you you Do Do verb verb you you Do Do verb verb fries fries want want you you noun noun with with fries fries want want prep prep that that with with fries fries pron pron that that with with pron pron want want you you Do Do
11 Properties of Sliding Windows Converts SSL to ordinary supervised learning Only captures the relationship between (part of) X and y t. Does not explicitly model relations among the y t s Assumes each window is independent
12 Recurrent Sliding Windows Recurrent Sliding Windows that that with with fries fries want want you you Do Do verb verb you you Do Do pron pron verb verb fries fries want want you you verb verb noun noun with with fries fries want want noun noun prep prep that that with with fries fries prep prep pron pron that that with with verb verb pron pron want want you you Do Do
13 Recurrent Sliding Windows Key Idea: Include y t as input feature when computing y t+. During training: Use the correct value of y t Or train iteratively (especially recurrent neural networks) During evaluation: Use the predicted value of y t
14 Properties of Recurrent Sliding Windows Captures relationship among the y s,, but only in one direction! Results on text-to to-speech: Method sliding window recurrent s. w. recurrent s. w. Direction none left-right right-left Words.5% 7.0%.% Letters 69.6% 67.9% 7.%
15 Hidden Markov Models Generalization of Naïve Bayes to SSL y y y y y 5 x x x x x 5 P(y ) P(y t y t- ) assumed the same for all t P(x t y t ) = P(x t, y t ) P(x t, y t ) L P(x assumed the same for all t P(x t,n,y,y t )
16 Making Predictions with HMMs Two possible goals: argmax Y P(Y X) find the most likely sequence of labels Y given the input sequence X argmax yt P(y t X) forall t find the most likely label y t at each time t given the entire input sequence X
17 Finding Most Likely Label Sequence: The Trellis s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Every label sequence corresponds to a path through the trellis graph. The probability of a label sequence is proportional to P(y ) P(x y ) P(y y ) P(x y ) L P(y T y T- ) P(x T y T )
18 Converting to Shortest Path Problem s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f max y,,y T P(y ) P(x y ) P(y y ) P(x y ) L P(y T y T- ) P(x T y T ) = min y,,y T l log [P(y ) P(x y )] + log [P(y y ) P(x y )] + L + log [P(y T y T- ) P(x T y T )] shortest path through graph. edge cost = log [P(y t y t- ) P(x t y t )]
19 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Step t of the Viterbi algorithm computes the possible successors of state y t- _ and computes the total path length for each edge
20 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Each node y t =k stores the cost µ of the shortest path that reaches it from s and the predecessor class y t- = k that achieves this cost k = argmin yt- log [P(y t y t- ) P(x t y t )] + µ(y t- ) µ(k) = min yt- log [P(y t y t- ) P(x t y t )] + µ(y t- )
21 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Compute Successors
22 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Compute and store shortest incoming arc at each node
23 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Compute successors
24 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Compute and store shortest incoming arc at each node
25 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Compute successors
26 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Compute and store shortest incoming edges
27 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Compute successors (trivial)
28 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Compute best edge into f
29 Finding Most Likely Label Sequence: The Viterbi Algorithm s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f Now trace back along best incoming edges to recover the predicted Y sequence: verb pronoun verb noun noun
30 Finding the Most Likely Label at time t: P(y t X) s verb pronoun verb noun noun verb pronoun noun adjective Do you want fries sir? f P(y = X) = probability of reaching y = from the start * probability of getting from y = to the finish
31 Finding the most likely class at each time t goal: compute P(y t x,, x T ) y:t :t- yt+:t P(y ) P(x y ) P(y y ) P(x y ) L P(y T y T- ) P(x T y T ) P(y y:t :t- ) P(x y ) P(y y ) P(x y ) L P(y t y t- ) P(x t y t ) yt+:t P(y t+ y t ) P(x t+ y t+ ) L P(y T y T- ) P(x T y T ) t+:t P(y yt- [ L y [ y P(y ) P(x y ) P(y y )] P(x y ) P(y t y t- )] P(x t y t ) P(y y )] L yt+ [P(y t+ y t ) P(x t+ y t+ ) L yt- [P(y T- y T- ) P(x T- y T- ) [P(y T y T- ) P(x T y T )]] L ]
32 Forward-Backward Algorithm α t (y t ) = yt- P(y t y t- ) P(x t y t ) α t- (y t- ) This is the sum over the arcs coming into y t = k It is computed forward along the sequence and stored in the trellis β t (y t ) = yt+ P(y t+ t+ y y t ) P( P(x t+ t+ y t+ ) β t+ t+ (y t+ ) It is computed backward along the sequence and stored in the trellis P(y t X) = α t (y t ) β t (y t ) / [ [ k α t (k) β t (k)]
33 Training Hidden Markov Models If the inputs and outputs are fully- observed, this is extremely easy: P(y =k) = [# examples with y =k] / m P(y t =k y t- = k ) k ) = [# <k,k > > transitions] / [# of times y t = k] P(x j = v y = k) = [# times y=k and x j =v ] / [# times y t = k] Should apply Laplace corrections to these estimates
34 Conditional Random Fields y y y x x x The y t s form a Markov Random Field conditioned on X: P(Y X) Lafferty, McCallum, & Pereira (00)
35 Markov Random Fields Graph G = (V,E) Each vertex v V represents a random variable y v. Each edge represents a direct probabilistic dependency. P(Y) = /Z exp [ [ c Ψ c (c(y))] c indexes the cliques in the graph Ψ c is a potential function c(y) ) selects the random variables participating in clique c.
36 A Simple MRF y y y Cliques: singletons: {y }, {y }, {y } pairs (edges); {y,y }, {y,y } P(hy,y,y i) ) = /Z exp[ψ (y ) + Ψ (y ) + Ψ (y ) + Ψ (y,y ) + Ψ (y,y )]
37 CRF Potential Functions are Conditioned on X y y y x x x Ψ t (y t,x): how compatible is y t with X? Ψ t,t- (y t,y t-,x): how compatible is a transition from y t- to y t with X?
38 CRF Potentials are Log Linear Models Ψ t (y t,x) ) = b β b g b (y t,x) Ψ t,t+ (y t,y t+,x) = a λ a f a (y t,y,y t+,x) where g b and f a are user-defined boolean functions ( features( features ) Example: g = [x[ t = o and y t = /@/]/] we will lump them together as Ψ t (y t, y t+,x) = a λ a f a (y t, y t+,x)
39 Making Predictions with CRFs Viterbi and Forward-Backward algorithms can be applied exactly as for HMMs
40 Training CRFs Let θ = {β{, β,, λ, λ, }} be all of our parameters Let F θ be our CRF, so F θ (Y,X) = P(Y X) Define the loss function L(Y,F θ (Y,X)) ) to be the Negative Log Likelihood L(Y,F θ (Y,X)) = log F θ (Y,X) Goal: Find θ to minimize loss (maximize likelihood) Algorithm: Gradient Descent
41 Gradient Computation gq = logp (Y X) λq = log λ q = λ q X = X t = X t t λ q X a Q t exp Ψ t(y t,y t,x) Z Ψ t (y t,y t,x) logz λafa(y t,y t,x) fq(y t,y t,x) λq logz λ q logz
42 Gradient of Z logz = Z λ q Z λ q = X Y exp Ψ t (yt 0 Z λq,y0 t,x) Y 0 t = X X exp Ψ t (yt 0 Z λq,y0 t,x) Y 0 t = X exp X Ψ t (yt 0 Z,y0 t,x) X Ψ t (yt 0 Y 0 t t λq,y0 t,x) h P = X exp t Ψ t(y 0 i t,y0 t,x) X X λ a f a (yt 0 Y 0 Z t λq,y0 t,x) a = X P (Y 0 X) fq(y 0 t,y0 t,x) Y 0 X t
43 Gradient Computation gq = X t fq(y t,y t,x) X Y 0 P (Y 0 X) X t fq(y 0 t,y 0 t,x) Number of times feature q is true minus the expected number of times feature q is true. This can be computed via the forward backward algorithm. First, apply forward-backward to compute P(y t-,y t X). P (y t,y t X) = Z X y t X y t α t (y t ) exp Ψ(y t,y t,x) β t (y t ) Then compute the gradient with respect to each λ q X g q = X f q (y t,y t,x) X t y t y t P (y t,y t X)f q (y t,y t,x)
44 Discriminative Methods Learn a discriminant function to which the Viterbi algorithm can be applied just get the right answer Methods: Averaged perceptron (Collins) Hidden Markov SVMs (Altun, et al.) Max Margin Markov Nets (Taskar, et al.)
45 Collins Perceptron Method If we ignore the global normalizer in the CRF, the score for a label sequence Y given an input sequence X is score(y ) = X t X a λ a f a (y t,y t,x) Collin s s approach is to adjust the weights λ a so that the correct label sequence gets the highest score according to the Viterbi algorithm
46 Sequence Perceptron Algorithm Initialize weights λ a = 0 For l =,,, L do For each training example (X i,y i ) apply Viterbi algorithm to find the path Ŷ with the highest score for all a, update λ a according to λ a := λ a + t [f a (y t,y t-,x) f a (ŷ t, ŷ t-, X)] Compares the viterbi path to the correct path.. Note that no update is made if the viterbi path is correct.
47 Averaged Perceptron Let λ l,i a be the value of λ a after processing training example i in iteration l Define λ * a = the average value of λ a = /(LN) l,i λ l,i a Use these averaged weights in the final classifier
48 Collins Part-of of-speech Tagging with Averaged Sequence Perceptron Without averaging:.68% error 0 iterations With averaging:.9% error 0 iterations
49 Hidden Markov SVM Define a kernel between two input values x and x : : k(x,x ). Define a kernel between (X,Y) and (X,Y,Y ) as follows: K((X,Y), (X,Y,Y )) = s,t I[y s- = y y t- & y s = y y t ] + I[y s = y y t ] k(x s,x t ) Number of (y t-,y t ) transitions that they share + Number of matching labels (weighted by similarity between the x values)
50 Dual Form of Linear Classifier Score(Y X) = j a α j (Y a ) K((X j,y a ), (X,Y)) a indexes support vector label sequences Y a Learning algorithm finds set of Y a label sequences weight values α j (Y a )
51 Dual Perceptron Algorithm Initialize α j = 0 For l from to L do For i from to N do Ŷ = argmax Y Score(Y X i ) if Ŷ Y i then α i (Y i ) = α i (Y i ) + α i (Ŷ)) = α i (Ŷ)
52 Hidden Markov SVM Algorithm For all i initialize S i = {Y i } set of support vector sequences for i α (Y)=0 i for all Y in S i For l from to L do For i from to N do Ŷ = argmax Y Yi Score(Y X i ) If Score(Y i X i ) < Score(Ŷ X i ) Add Ŷ to S i Solve quadratic program to optimize the α i (Y) for all Y in S i to maximize the margin between Y i and all of the other Y s Y s in S i If α i (Y) = 0, delete Y from S i
53 Altun et al. comparison
54 Maximum Margin Markov Networks Define SVM-like optimization problem to maximize the per time step margin Define F(X i,y i,ŷ)) = F(X i,y i ) F(X i,ŷ) Y(Y i, Ŷ) ) = i I[ŷ t y it ] MMM SVM formulation: min w + C i ξ i subject to w F(X i,y i,ŷ) Y(Y i, Ŷ) ) + ξ i forall Y, forall i
55 Dual Form maximize i Y α i (Ŷ) (Y i, Ŷ) ½ i Ŷ j Ŷ α i (Ŷ) α j (Ŷ )) [ F(X[ i,y i,ŷ) F(X j,y j,ŷ )] subject to Ŷ α i (Ŷ)) = C forall i α i (Ŷ) 0 forall i, forall Ŷ Note that there are exponentially-many Ŷ label sequences
56 Converting to a Polynomial-Sized Note the constraints: Ŷ α i (Ŷ)) = C forall i Formulation α i (Ŷ) 0 forall i, forall Ŷ These imply that for each i, the α i (Ŷ)) values are proportional to a probability distribution: Q(Ŷ X i ) = α i (Ŷ)) / C Because the MRF is a simple chain, this distribution can be factored into local distributions: Q(Ŷ X i ) = t Q(ŷ t-, ŷ t X i ) Let µ i (ŷ t-, ŷ t ) be the unnormalized version of Q
57 Reformulated Dual Form max X i X t X i,j X ŷ t µ i (ŷ t )I[ŷ t 6= y i,t ] X t X X X ŷ t,ŷ t s ŷ 0 s,ŷ 0 s µ i (ŷ t, ŷ t )µ j (ŷ 0 s, ŷ 0 s) F (ŷ t, ŷ t,x i ) F (ŷ 0 s, ŷ0 s,x j ) subject to X µ i (ŷ t, ŷ t ) = µ i (ŷ t ) ŷ t X µ i (ŷ t ) = C ŷ t µ i (ŷ t, ŷ t ) 0
58 Variables in the Dual Form µ i (k,k ) ) for each training example i and each possible class labels k, k : k : O(NK ) µ i (k) for each trianing example i and possible class label k: O(NK) Polynomial!
59 Taskar et al. comparison Handwriting Recognition log-reg: logistic regression sliding window CRF: msvm: multiclass SVM sliding window M^N: max margin markov net
60 Current State of the Art Discriminative Methods give best results not clear whether they scale published results all involve small numbers of training examples and very long training times Work is continuing on making CRFs fast and practical new methods for training CRFs potentially extendable to discriminative methods
Conditional Random Fields for Sequential Supervised Learning
Conditional Random Fields for Sequential Supervised Learning Thomas G. Dietterich Adam Ashenfelter Department of Computer Science Oregon State University Corvallis, Oregon 97331 http://www.eecs.oregonstate.edu/~tgd
More informationSequence labeling. Taking collective a set of interrelated instances x 1,, x T and jointly labeling them
HMM, MEMM and CRF 40-957 Special opics in Artificial Intelligence: Probabilistic Graphical Models Sharif University of echnology Soleymani Spring 2014 Sequence labeling aking collective a set of interrelated
More informationConditional Random Field
Introduction Linear-Chain General Specific Implementations Conclusions Corso di Elaborazione del Linguaggio Naturale Pisa, May, 2011 Introduction Linear-Chain General Specific Implementations Conclusions
More informationMachine Learning for Structured Prediction
Machine Learning for Structured Prediction Grzegorz Chrupa la National Centre for Language Technology School of Computing Dublin City University NCLT Seminar Grzegorz Chrupa la (DCU) Machine Learning for
More informationUndirected Graphical Models
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional
More informationStatistical Machine Learning Theory. From Multi-class Classification to Structured Output Prediction. Hisashi Kashima.
http://goo.gl/jv7vj9 Course website KYOTO UNIVERSITY Statistical Machine Learning Theory From Multi-class Classification to Structured Output Prediction Hisashi Kashima kashima@i.kyoto-u.ac.jp DEPARTMENT
More informationStatistical Machine Learning Theory. From Multi-class Classification to Structured Output Prediction. Hisashi Kashima.
http://goo.gl/xilnmn Course website KYOTO UNIVERSITY Statistical Machine Learning Theory From Multi-class Classification to Structured Output Prediction Hisashi Kashima kashima@i.kyoto-u.ac.jp DEPARTMENT
More informationSequence Modelling with Features: Linear-Chain Conditional Random Fields. COMP-599 Oct 6, 2015
Sequence Modelling with Features: Linear-Chain Conditional Random Fields COMP-599 Oct 6, 2015 Announcement A2 is out. Due Oct 20 at 1pm. 2 Outline Hidden Markov models: shortcomings Generative vs. discriminative
More informationStatistical NLP for the Web Log Linear Models, MEMM, Conditional Random Fields
Statistical NLP for the Web Log Linear Models, MEMM, Conditional Random Fields Sameer Maskey Week 13, Nov 28, 2012 1 Announcements Next lecture is the last lecture Wrap up of the semester 2 Final Project
More informationHidden Markov Models
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Hidden Markov Models Matt Gormley Lecture 19 Nov. 5, 2018 1 Reminders Homework
More informationA brief introduction to Conditional Random Fields
A brief introduction to Conditional Random Fields Mark Johnson Macquarie University April, 2005, updated October 2010 1 Talk outline Graphical models Maximum likelihood and maximum conditional likelihood
More informationHidden Markov Models
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Hidden Markov Models Matt Gormley Lecture 22 April 2, 2018 1 Reminders Homework
More information10 : HMM and CRF. 1 Case Study: Supervised Part-of-Speech Tagging
10-708: Probabilistic Graphical Models 10-708, Spring 2018 10 : HMM and CRF Lecturer: Kayhan Batmanghelich Scribes: Ben Lengerich, Michael Kleyman 1 Case Study: Supervised Part-of-Speech Tagging We will
More informationConditional Random Fields and beyond DANIEL KHASHABI CS 546 UIUC, 2013
Conditional Random Fields and beyond DANIEL KHASHABI CS 546 UIUC, 2013 Outline Modeling Inference Training Applications Outline Modeling Problem definition Discriminative vs. Generative Chain CRF General
More informationGraphical models for part of speech tagging
Indian Institute of Technology, Bombay and Research Division, India Research Lab Graphical models for part of speech tagging Different Models for POS tagging HMM Maximum Entropy Markov Models Conditional
More informationLecture 13: Structured Prediction
Lecture 13: Structured Prediction Kai-Wei Chang CS @ University of Virginia kw@kwchang.net Couse webpage: http://kwchang.net/teaching/nlp16 CS6501: NLP 1 Quiz 2 v Lectures 9-13 v Lecture 12: before page
More informationLecture 9: PGM Learning
13 Oct 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Table of Contents I Learning parameters in MRFs 1 Learning parameters in MRFs Inference and Learning Given parameters (of potentials) and
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 24, 2016 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationAlgorithms for Predicting Structured Data
1 / 70 Algorithms for Predicting Structured Data Thomas Gärtner / Shankar Vembu Fraunhofer IAIS / UIUC ECML PKDD 2010 Structured Prediction 2 / 70 Predicting multiple outputs with complex internal structure
More informationProbabilistic Graphical Models
School of Computer Science Probabilistic Graphical Models Max-margin learning of GM Eric Xing Lecture 28, Apr 28, 2014 b r a c e Reading: 1 Classical Predictive Models Input and output space: Predictive
More informationProbabilistic Models for Sequence Labeling
Probabilistic Models for Sequence Labeling Besnik Fetahu June 9, 2011 Besnik Fetahu () Probabilistic Models for Sequence Labeling June 9, 2011 1 / 26 Background & Motivation Problem introduction Generative
More informationLogistic Regression: Online, Lazy, Kernelized, Sequential, etc.
Logistic Regression: Online, Lazy, Kernelized, Sequential, etc. Harsha Veeramachaneni Thomson Reuter Research and Development April 1, 2010 Harsha Veeramachaneni (TR R&D) Logistic Regression April 1, 2010
More informationA gentle introduction to Hidden Markov Models
A gentle introduction to Hidden Markov Models Mark Johnson Brown University November 2009 1 / 27 Outline What is sequence labeling? Markov models Hidden Markov models Finding the most likely state sequence
More informationInformation Extraction from Text
Information Extraction from Text Jing Jiang Chapter 2 from Mining Text Data (2012) Presented by Andrew Landgraf, September 13, 2013 1 What is Information Extraction? Goal is to discover structured information
More informationPartially Directed Graphs and Conditional Random Fields. Sargur Srihari
Partially Directed Graphs and Conditional Random Fields Sargur srihari@cedar.buffalo.edu 1 Topics Conditional Random Fields Gibbs distribution and CRF Directed and Undirected Independencies View as combination
More informationLogistic Regression & Neural Networks
Logistic Regression & Neural Networks CMSC 723 / LING 723 / INST 725 Marine Carpuat Slides credit: Graham Neubig, Jacob Eisenstein Logistic Regression Perceptron & Probabilities What if we want a probability
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 19 Oct, 23, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models D. Page,
More informationSequence Labeling: HMMs & Structured Perceptron
Sequence Labeling: HMMs & Structured Perceptron CMSC 723 / LING 723 / INST 725 MARINE CARPUAT marine@cs.umd.edu HMM: Formal Specification Q: a finite set of N states Q = {q 0, q 1, q 2, q 3, } N N Transition
More informationStructured Prediction
Structured Prediction Classification Algorithms Classify objects x X into labels y Y First there was binary: Y = {0, 1} Then multiclass: Y = {1,...,6} The next generation: Structured Labels Structured
More informationPredicting Sequences: Structured Perceptron. CS 6355: Structured Prediction
Predicting Sequences: Structured Perceptron CS 6355: Structured Prediction 1 Conditional Random Fields summary An undirected graphical model Decompose the score over the structure into a collection of
More informationHidden Markov Models
CS769 Spring 2010 Advanced Natural Language Processing Hidden Markov Models Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu 1 Part-of-Speech Tagging The goal of Part-of-Speech (POS) tagging is to label each
More informationHidden Markov Models in Language Processing
Hidden Markov Models in Language Processing Dustin Hillard Lecture notes courtesy of Prof. Mari Ostendorf Outline Review of Markov models What is an HMM? Examples General idea of hidden variables: implications
More informationCS446: Machine Learning Fall Final Exam. December 6 th, 2016
CS446: Machine Learning Fall 2016 Final Exam December 6 th, 2016 This is a closed book exam. Everything you need in order to solve the problems is supplied in the body of this exam. This exam booklet contains
More informationPart of Speech Tagging: Viterbi, Forward, Backward, Forward- Backward, Baum-Welch. COMP-599 Oct 1, 2015
Part of Speech Tagging: Viterbi, Forward, Backward, Forward- Backward, Baum-Welch COMP-599 Oct 1, 2015 Announcements Research skills workshop today 3pm-4:30pm Schulich Library room 313 Start thinking about
More informationBayesian Networks Introduction to Machine Learning. Matt Gormley Lecture 24 April 9, 2018
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Bayesian Networks Matt Gormley Lecture 24 April 9, 2018 1 Homework 7: HMMs Reminders
More informationStatistical Methods for NLP
Statistical Methods for NLP Information Extraction, Hidden Markov Models Sameer Maskey Week 5, Oct 3, 2012 *many slides provided by Bhuvana Ramabhadran, Stanley Chen, Michael Picheny Speech Recognition
More information2.2 Structured Prediction
The hinge loss (also called the margin loss), which is optimized by the SVM, is a ramp function that has slope 1 when yf(x) < 1 and is zero otherwise. Two other loss functions squared loss and exponential
More informationHidden Markov Models
Hidden Markov Models CI/CI(CS) UE, SS 2015 Christian Knoll Signal Processing and Speech Communication Laboratory Graz University of Technology June 23, 2015 CI/CI(CS) SS 2015 June 23, 2015 Slide 1/26 Content
More informationHidden Markov Models
Hidden Markov Models Slides mostly from Mitch Marcus and Eric Fosler (with lots of modifications). Have you seen HMMs? Have you seen Kalman filters? Have you seen dynamic programming? HMMs are dynamic
More informationLecture 4: Hidden Markov Models: An Introduction to Dynamic Decision Making. November 11, 2010
Hidden Lecture 4: Hidden : An Introduction to Dynamic Decision Making November 11, 2010 Special Meeting 1/26 Markov Model Hidden When a dynamical system is probabilistic it may be determined by the transition
More informationLog-Linear Models, MEMMs, and CRFs
Log-Linear Models, MEMMs, and CRFs Michael Collins 1 Notation Throughout this note I ll use underline to denote vectors. For example, w R d will be a vector with components w 1, w 2,... w d. We use expx
More informationGeneralized Linear Classifiers in NLP
Generalized Linear Classifiers in NLP (or Discriminative Generalized Linear Feature-Based Classifiers) Graduate School of Language Technology, Sweden 2009 Ryan McDonald Google Inc., New York, USA E-mail:
More informationWhat s an HMM? Extraction with Finite State Machines e.g. Hidden Markov Models (HMMs) Hidden Markov Models (HMMs) for Information Extraction
Hidden Markov Models (HMMs) for Information Extraction Daniel S. Weld CSE 454 Extraction with Finite State Machines e.g. Hidden Markov Models (HMMs) standard sequence model in genomics, speech, NLP, What
More informationRandom Field Models for Applications in Computer Vision
Random Field Models for Applications in Computer Vision Nazre Batool Post-doctorate Fellow, Team AYIN, INRIA Sophia Antipolis Outline Graphical Models Generative vs. Discriminative Classifiers Markov Random
More informationMachine Learning for NLP
Machine Learning for NLP Linear Models Joakim Nivre Uppsala University Department of Linguistics and Philology Slides adapted from Ryan McDonald, Google Research Machine Learning for NLP 1(26) Outline
More informationCS838-1 Advanced NLP: Hidden Markov Models
CS838-1 Advanced NLP: Hidden Markov Models Xiaojin Zhu 2007 Send comments to jerryzhu@cs.wisc.edu 1 Part of Speech Tagging Tag each word in a sentence with its part-of-speech, e.g., The/AT representative/nn
More informationActive learning in sequence labeling
Active learning in sequence labeling Tomáš Šabata 11. 5. 2017 Czech Technical University in Prague Faculty of Information technology Department of Theoretical Computer Science Table of contents 1. Introduction
More informationMachine Learning Practice Page 2 of 2 10/28/13
Machine Learning 10-701 Practice Page 2 of 2 10/28/13 1. True or False Please give an explanation for your answer, this is worth 1 pt/question. (a) (2 points) No classifier can do better than a naive Bayes
More informationLecture 13: Discriminative Sequence Models (MEMM and Struct. Perceptron)
Lecture 13: Discriminative Sequence Models (MEMM and Struct. Perceptron) Intro to NLP, CS585, Fall 2014 http://people.cs.umass.edu/~brenocon/inlp2014/ Brendan O Connor (http://brenocon.com) 1 Models for
More informationAn Introduction to Machine Learning
An Introduction to Machine Learning L6: Structured Estimation Alexander J. Smola Statistical Machine Learning Program Canberra, ACT 0200 Australia Alex.Smola@nicta.com.au Tata Institute, Pune, January
More informationIntroduction to Logistic Regression and Support Vector Machine
Introduction to Logistic Regression and Support Vector Machine guest lecturer: Ming-Wei Chang CS 446 Fall, 2009 () / 25 Fall, 2009 / 25 Before we start () 2 / 25 Fall, 2009 2 / 25 Before we start Feel
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationSupport Vector Machines
Support Vector Machines Le Song Machine Learning I CSE 6740, Fall 2013 Naïve Bayes classifier Still use Bayes decision rule for classification P y x = P x y P y P x But assume p x y = 1 is fully factorized
More informationMulticlass and Introduction to Structured Prediction
Multiclass and Introduction to Structured Prediction David S. Rosenberg New York University March 27, 2018 David S. Rosenberg (New York University) DS-GA 1003 / CSCI-GA 2567 March 27, 2018 1 / 49 Contents
More informationCOMP90051 Statistical Machine Learning
COMP90051 Statistical Machine Learning Semester 2, 2017 Lecturer: Trevor Cohn 24. Hidden Markov Models & message passing Looking back Representation of joint distributions Conditional/marginal independence
More informationMidterm sample questions
Midterm sample questions CS 585, Brendan O Connor and David Belanger October 12, 2014 1 Topics on the midterm Language concepts Translation issues: word order, multiword translations Human evaluation Parts
More informationCS534: Machine Learning. Thomas G. Dietterich 221C Dearborn Hall
CS534: Machine Learning Thomas G. Dietterich 221C Dearborn Hall tgd@cs.orst.edu http://www.cs.orst.edu/~tgd/classes/534 1 Course Overview Introduction: Basic problems and questions in machine learning.
More informationMulticlass and Introduction to Structured Prediction
Multiclass and Introduction to Structured Prediction David S. Rosenberg Bloomberg ML EDU November 28, 2017 David S. Rosenberg (Bloomberg ML EDU) ML 101 November 28, 2017 1 / 48 Introduction David S. Rosenberg
More informationlecture 6: modeling sequences (final part)
Natural Language Processing 1 lecture 6: modeling sequences (final part) Ivan Titov Institute for Logic, Language and Computation Outline After a recap: } Few more words about unsupervised estimation of
More informationIntelligent Systems (AI-2)
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 18 Oct, 21, 2015 Slide Sources Raymond J. Mooney University of Texas at Austin D. Koller, Stanford CS - Probabilistic Graphical Models CPSC
More informationMACHINE LEARNING FOR NATURAL LANGUAGE PROCESSING
MACHINE LEARNING FOR NATURAL LANGUAGE PROCESSING Outline Some Sample NLP Task [Noah Smith] Structured Prediction For NLP Structured Prediction Methods Conditional Random Fields Structured Perceptron Discussion
More informationConditional Random Fields: An Introduction
University of Pennsylvania ScholarlyCommons Technical Reports (CIS) Department of Computer & Information Science 2-24-2004 Conditional Random Fields: An Introduction Hanna M. Wallach University of Pennsylvania
More informationStatistical Methods for NLP
Statistical Methods for NLP Text Categorization, Support Vector Machines Sameer Maskey Announcement Reading Assignments Will be posted online tonight Homework 1 Assigned and available from the course website
More informationCRF for human beings
CRF for human beings Arne Skjærholt LNS seminar CRF for human beings LNS seminar 1 / 29 Let G = (V, E) be a graph such that Y = (Y v ) v V, so that Y is indexed by the vertices of G. Then (X, Y) is a conditional
More informationGraphical Models for Collaborative Filtering
Graphical Models for Collaborative Filtering Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Sequence modeling HMM, Kalman Filter, etc.: Similarity: the same graphical model topology,
More informationCSCI-567: Machine Learning (Spring 2019)
CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Mar. 19, 2019 March 19, 2019 1 / 43 Administration March 19, 2019 2 / 43 Administration TA3 is due this week March
More informationp(d θ ) l(θ ) 1.2 x x x
p(d θ ).2 x 0-7 0.8 x 0-7 0.4 x 0-7 l(θ ) -20-40 -60-80 -00 2 3 4 5 6 7 θ ˆ 2 3 4 5 6 7 θ ˆ 2 3 4 5 6 7 θ θ x FIGURE 3.. The top graph shows several training points in one dimension, known or assumed to
More informationMidterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas
Midterm Review CS 6375: Machine Learning Vibhav Gogate The University of Texas at Dallas Machine Learning Supervised Learning Unsupervised Learning Reinforcement Learning Parametric Y Continuous Non-parametric
More informationMachine Learning Basics Lecture 2: Linear Classification. Princeton University COS 495 Instructor: Yingyu Liang
Machine Learning Basics Lecture 2: Linear Classification Princeton University COS 495 Instructor: Yingyu Liang Review: machine learning basics Math formulation Given training data x i, y i : 1 i n i.i.d.
More informationProbabilistic classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2016
Probabilistic classification CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2016 Topics Probabilistic approach Bayes decision theory Generative models Gaussian Bayes classifier
More informationMore on HMMs and other sequence models. Intro to NLP - ETHZ - 18/03/2013
More on HMMs and other sequence models Intro to NLP - ETHZ - 18/03/2013 Summary Parts of speech tagging HMMs: Unsupervised parameter estimation Forward Backward algorithm Bayesian variants Discriminative
More informationMachine Learning for NLP
Machine Learning for NLP Uppsala University Department of Linguistics and Philology Slides borrowed from Ryan McDonald, Google Research Machine Learning for NLP 1(50) Introduction Linear Classifiers Classifiers
More informationFINAL: CS 6375 (Machine Learning) Fall 2014
FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for
More informationStatistical Data Mining and Machine Learning Hilary Term 2016
Statistical Data Mining and Machine Learning Hilary Term 2016 Dino Sejdinovic Department of Statistics Oxford Slides and other materials available at: http://www.stats.ox.ac.uk/~sejdinov/sdmml Naïve Bayes
More informationExpectation Maximization (EM)
Expectation Maximization (EM) The EM algorithm is used to train models involving latent variables using training data in which the latent variables are not observed (unlabeled data). This is to be contrasted
More informationParameter learning in CRF s
Parameter learning in CRF s June 01, 2009 Structured output learning We ish to learn a discriminant (or compatability) function: F : X Y R (1) here X is the space of inputs and Y is the space of outputs.
More informationGraphical Models Seminar
Graphical Models Seminar Forward-Backward and Viterbi Algorithm for HMMs Bishop, PRML, Chapters 13.2.2, 13.2.3, 13.2.5 Dinu Kaufmann Departement Mathematik und Informatik Universität Basel April 8, 2013
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table
More informationMachine Learning for Signal Processing Bayes Classification and Regression
Machine Learning for Signal Processing Bayes Classification and Regression Instructor: Bhiksha Raj 11755/18797 1 Recap: KNN A very effective and simple way of performing classification Simple model: For
More informationHidden Markov Models Part 2: Algorithms
Hidden Markov Models Part 2: Algorithms CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Hidden Markov Model An HMM consists of:
More informationMachine Learning. Classification, Discriminative learning. Marc Toussaint University of Stuttgart Summer 2015
Machine Learning Classification, Discriminative learning Structured output, structured input, discriminative function, joint input-output features, Likelihood Maximization, Logistic regression, binary
More informationGenerative v. Discriminative classifiers Intuition
Logistic Regression Machine Learning 070/578 Carlos Guestrin Carnegie Mellon University September 24 th, 2007 Generative v. Discriminative classifiers Intuition Want to Learn: h:x a Y X features Y target
More information1 Machine Learning Concepts (16 points)
CSCI 567 Fall 2018 Midterm Exam DO NOT OPEN EXAM UNTIL INSTRUCTED TO DO SO PLEASE TURN OFF ALL CELL PHONES Problem 1 2 3 4 5 6 Total Max 16 10 16 42 24 12 120 Points Please read the following instructions
More informationKernel methods, kernel SVM and ridge regression
Kernel methods, kernel SVM and ridge regression Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Collaborative Filtering 2 Collaborative Filtering R: rating matrix; U: user factor;
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More informationAdvanced Data Science
Advanced Data Science Dr. Kira Radinsky Slides Adapted from Tom M. Mitchell Agenda Topics Covered: Time series data Markov Models Hidden Markov Models Dynamic Bayes Nets Additional Reading: Bishop: Chapter
More informationLinear Classifiers IV
Universität Potsdam Institut für Informatik Lehrstuhl Linear Classifiers IV Blaine Nelson, Tobias Scheffer Contents Classification Problem Bayesian Classifier Decision Linear Classifiers, MAP Models Logistic
More informationDiscriminative Models
No.5 Discriminative Models Hui Jiang Department of Electrical Engineering and Computer Science Lassonde School of Engineering York University, Toronto, Canada Outline Generative vs. Discriminative models
More informationTopics in Structured Prediction: Problems and Approaches
Topics in Structured Prediction: Problems and Approaches DRAFT Ankan Saha June 9, 2010 Abstract We consider the task of structured data prediction. Over the last few years, there has been an abundance
More informationBrief Introduction of Machine Learning Techniques for Content Analysis
1 Brief Introduction of Machine Learning Techniques for Content Analysis Wei-Ta Chu 2008/11/20 Outline 2 Overview Gaussian Mixture Model (GMM) Hidden Markov Model (HMM) Support Vector Machine (SVM) Overview
More informationMachine Learning & Data Mining Caltech CS/CNS/EE 155 Hidden Markov Models Last Updated: Feb 7th, 2017
1 Introduction Let x = (x 1,..., x M ) denote a sequence (e.g. a sequence of words), and let y = (y 1,..., y M ) denote a corresponding hidden sequence that we believe explains or influences x somehow
More informationProbabilistic Graphical Models: MRFs and CRFs. CSE628: Natural Language Processing Guest Lecturer: Veselin Stoyanov
Probabilistic Graphical Models: MRFs and CRFs CSE628: Natural Language Processing Guest Lecturer: Veselin Stoyanov Why PGMs? PGMs can model joint probabilities of many events. many techniques commonly
More informationMachine Learning. Regression-Based Classification & Gaussian Discriminant Analysis. Manfred Huber
Machine Learning Regression-Based Classification & Gaussian Discriminant Analysis Manfred Huber 2015 1 Logistic Regression Linear regression provides a nice representation and an efficient solution to
More informationDiscriminative Models
No.5 Discriminative Models Hui Jiang Department of Electrical Engineering and Computer Science Lassonde School of Engineering York University, Toronto, Canada Outline Generative vs. Discriminative models
More informationMachine Learning Basics Lecture 7: Multiclass Classification. Princeton University COS 495 Instructor: Yingyu Liang
Machine Learning Basics Lecture 7: Multiclass Classification Princeton University COS 495 Instructor: Yingyu Liang Example: image classification indoor Indoor outdoor Example: image classification (multiclass)
More informationLecture 15. Probabilistic Models on Graph
Lecture 15. Probabilistic Models on Graph Prof. Alan Yuille Spring 2014 1 Introduction We discuss how to define probabilistic models that use richly structured probability distributions and describe how
More informationPattern Recognition and Machine Learning. Bishop Chapter 6: Kernel Methods
Pattern Recognition and Machine Learning Chapter 6: Kernel Methods Vasil Khalidov Alex Kläser December 13, 2007 Training Data: Keep or Discard? Parametric methods (linear/nonlinear) so far: learn parameter
More informationwith Local Dependencies
CS11-747 Neural Networks for NLP Structured Prediction with Local Dependencies Xuezhe Ma (Max) Site https://phontron.com/class/nn4nlp2017/ An Example Structured Prediction Problem: Sequence Labeling Sequence
More informationNaïve Bayes classification
Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss
More informationDirected Probabilistic Graphical Models CMSC 678 UMBC
Directed Probabilistic Graphical Models CMSC 678 UMBC Announcement 1: Assignment 3 Due Wednesday April 11 th, 11:59 AM Any questions? Announcement 2: Progress Report on Project Due Monday April 16 th,
More information