Statistical NLP Spring A Discriminative Approach

Similar documents
Natural Language Processing. Classification. Features. Some Definitions. Classification. Feature Vectors. Classification I. Dan Klein UC Berkeley

Statistical NLP Spring 2010

Classification. Statistical NLP Spring Example: Text Classification. Some Definitions. Block Feature Vectors.

Natural Language Processing. Classification. Margin. Linear Models: Perceptron. Issues with Perceptrons. Problems with Perceptrons.

Algorithms for NLP. Classification II. Taylor Berg-Kirkpatrick CMU Slides: Dan Klein UC Berkeley

CS 188: Artificial Intelligence. Outline

EE 511 Online Learning Perceptron

Learning: Binary Perceptron. Examples: Perceptron. Separable Case. In the space of feature vectors

Text Classification. CS5740: Natural Language Processing Spring Instructor: Yoav Artzi

CS 188: Artificial Intelligence Fall 2008

General Naïve Bayes. CS 188: Artificial Intelligence Fall Example: Overfitting. Example: OCR. Example: Spam Filtering. Example: Spam Filtering

CS 188: Artificial Intelligence Fall 2011

Errors, and What to Do. CS 188: Artificial Intelligence Fall What to Do About Errors. Later On. Some (Simplified) Biology

MIRA, SVM, k-nn. Lirong Xia

Machine Learning for NLP

CS 188: Artificial Intelligence. Machine Learning

Machine Learning for NLP

Support vector machines Lecture 4

CS 343: Artificial Intelligence

CSEP%517% Natural%Language%Processing% Autumn%2013%

Introduction to Logistic Regression and Support Vector Machine

Machine Learning. Hal Daumé III. Computer Science University of Maryland CS 421: Introduction to Artificial Intelligence 8 May 2012

CS 5522: Artificial Intelligence II

From Binary to Multiclass Classification. CS 6961: Structured Prediction Spring 2018

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

Algorithms for NLP. Classifica(on III. Taylor Berg- Kirkpatrick CMU Slides: Dan Klein UC Berkeley

Stochastic Gradient Descent

Support Vector Machines

Perceptron. Subhransu Maji. CMPSCI 689: Machine Learning. 3 February February 2015

CPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017

Logistic Regression Logistic

Logistic Regression. Machine Learning Fall 2018

Max Margin-Classifier

Lecture 4: Training a Classifier

Lecture 9: Large Margin Classifiers. Linear Support Vector Machines

Ad Placement Strategies

Case Study 1: Estimating Click Probabilities. Kakade Announcements: Project Proposals: due this Friday!

Machine Learning for Structured Prediction

Gaussian and Linear Discriminant Analysis; Multiclass Classification

CS6375: Machine Learning Gautam Kunapuli. Support Vector Machines

ML4NLP Multiclass Classification

Machine Learning: Chenhao Tan University of Colorado Boulder LECTURE 9

Classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

Announcements - Homework

Introduction to Machine Learning

Algorithms for NLP. Language Modeling III. Taylor Berg-Kirkpatrick CMU Slides: Dan Klein UC Berkeley

CSC 411 Lecture 17: Support Vector Machine

CPSC 340: Machine Learning and Data Mining

Applied Machine Learning Lecture 5: Linear classifiers, continued. Richard Johansson

Machine Learning Lecture 5

Linear & nonlinear classifiers

18.9 SUPPORT VECTOR MACHINES

CS 188: Artificial Intelligence Spring Today

Naïve Bayes, Maxent and Neural Models

Linear Classifiers and the Perceptron

SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning

Logistic Regression. Some slides adapted from Dan Jurfasky and Brendan O Connor

Logistic Regression & Neural Networks

Lecture 4: Training a Classifier

Perceptron (Theory) + Linear Regression

Kernel Methods and Support Vector Machines

Multiclass Classification-1

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Optimization and Gradient Descent

Linear Models for Classification: Discriminative Learning (Perceptron, SVMs, MaxEnt)

Midterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas

Deep Learning for Computer Vision

Warm up: risk prediction with logistic regression

CS 446 Machine Learning Fall 2016 Nov 01, Bayesian Learning

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

Linear discriminant functions

Lecture 3: Multiclass Classification

Linear smoother. ŷ = S y. where s ij = s ij (x) e.g. s ij = diag(l i (x))

ECE521 Lecture7. Logistic Regression

Jeff Howbert Introduction to Machine Learning Winter

CS6375: Machine Learning Gautam Kunapuli. Decision Trees

A short introduction to supervised learning, with applications to cancer pathway analysis Dr. Christina Leslie

Midterm Review CS 7301: Advanced Machine Learning. Vibhav Gogate The University of Texas at Dallas

Machine Learning And Applications: Supervised Learning-SVM

Classification Logistic Regression

Logistic Regression. COMP 527 Danushka Bollegala

Ad Placement Strategies

Logistic Regression Introduction to Machine Learning. Matt Gormley Lecture 9 Sep. 26, 2018

CSE446: Clustering and EM Spring 2017

Lecture 14 : Online Learning, Stochastic Gradient Descent, Perceptron

Classification II: Decision Trees and SVMs

Dan Roth 461C, 3401 Walnut

ECE521 week 3: 23/26 January 2017

Classification and Regression Trees

Support Vector Machines

Minimax risk bounds for linear threshold functions

Introduction to AI Learning Bayesian networks. Vibhav Gogate

Math for Machine Learning Open Doors to Data Science and Artificial Intelligence. Richard Han

MIDTERM SOLUTIONS: FALL 2012 CS 6375 INSTRUCTOR: VIBHAV GOGATE

Brief Introduction to Machine Learning

Support Vector Machines

Midterm. Introduction to Machine Learning. CS 189 Spring Please do not open the exam before you are instructed to do so.

Machine Learning Lecture 7

Machine Learning Practice Page 2 of 2 10/28/13

Transcription:

Statistical NLP Spring 2008 Lecture 6: Classification Dan Klein UC Berkeley A Discriminative Approach View WSD as a discrimination task (regression, really) P(sense context:jail, context:county, context:feeding, local-context:jail, local-context:meals subcat:np, direct-object-head:meals,.) Have to estimate multinomial (over senses) where there are a huge number of things to condition on History is too complex to think about this as a smoothing / backoff problem Many feature-based classification techniques out there We tend to need ones that output distributions over classes (why?) 1

Feature Representations d Washington County jail served 11,166 meals last month - a figure that translates to feeding some 120 people three times daily for 31 days. Features are indicator functions f i which count the occurrences of certain patterns in the input We map each input to a vector of feature predicate counts { fi( d)} context:jail = 1 context:county = 1 context:feeding = 1 context:game = 0 local-context:jail = 1 local-context:meals = 1 subcat:np = 1 subcat:pp = 0 object-head:meals = 1 object-head:ball = 0 Example: Text Classification We want to classify documents into categories DOCUMENT win the election win the game see a movie CATEGORY POLITICS SPORTS OTHER Classically, do this on the basis of words in the document, but other information sources are potentially relevant: Document length Average word length Document s source Document layout 2

Some Definitions Sometimes, we want Y to depend on x INPUTS OUTPUT SPACE OUTPUTS TRUE OUTPUTS FEATURE VECTORS Either x is implicit, or y contains x win the election SPORTS, POLITICS, OTHER SPORTS POLITICS SPORTS win POLITICS election POLITICS win Block Feature Vectors Sometimes, we think of the input as having features, which are multiplied by outputs to form the candidates win the election win election 3

Non-Block Feature Vectors Sometimes the features of candidates cannot be decomposed in this regular way Example: a parse tree s features may be the productions present in the tree N NP NP N S S VP VP V N V N Different candidates will thus often share features We ll return to the non-block case later NP N V S NP VP V N NP N VP VP N Linear Models: Scoring In a linear model, each feature gets a weight w We compare hypotheses on the basis of their linear scores: 4

Linear Models: Prediction Rule The linear prediction rule: We ve said nothing about where weights come from! Binary Decision Rule Heavily studied case: binary classification Simplifed: only class 1 has features Decision rule is a hyperplane One side will be class 1 Other side will be class 0 money 2 SPAM 1 BIAS : -3 free : 4 money : 2 the : 0... HAM 0 0 1 free 5

Multiclass Decision Rule If more than two classes: Highest score wins Boundaries are more complex Harder to visualize There are other ways: e.g. reconcile pairwise decisions Learning Classifier Weights Two broad approaches to learning weights Generative: work with a probabilistic model of the data, weights are (log) local conditional probabilities Advantages: learning weights is easy, smoothing is wellunderstood, backed by understanding of modeling Discriminative: set weights based on some errorrelated criterion Advantages: error-driven, often weights which are good for classification aren t the ones which best describe the data We ll mainly talk about the latter 6

Example: Stoplights Lights Working Reality Lights Broken P(g,r,w) = 3/7 P(r,g,w) = 3/7 P(r,r,b) = 1/7 NB Model Working? NS EW NB FACTORS: P(w) = 6/7 P(r w) = 1/2 P(g w) = 1/2 P(b) = 1/7 P(r b) = 1 P(g b) = 0 Example: Stoplights What does the model say when both lights are red? P(b,r,r) = (1/7)(1)(1) = 1/7 = 4/28 P(w,r,r) = (6/7)(1/2)(1/2) = 6/28 = 6/28 P(w r,r) = 6/10! We ll guess that (r,r) indicates lights are working Imagine if P(b) were boosted higher, to 1/2: P(b,r,r) = (1/2)(1)(1) = 1/2 = 4/8 P(w,r,r) = (1/2)(1/2)(1/2) = 1/8 = 1/8 P(w r,r) = 1/5! Non-generative values can give better classification 7

Linear Models: Naïve-Bayes (Multinomial) Naïve-Bayes is a linear model, where: y d 1 d 2 d n How to pick weights? Goal: choose best vector w given training data For now, we mean best for classification The ideal: the weights which have greatest test set accuracy / F1 / whatever But, don t have the test set Must compute weights from training set Maybe we want weights which give best training set accuracy? Hard discontinuous optimization problem May not (does not) generalize to test set Easy to overfit Though, min-error training for MT does exactly this. 8

Minimize Training Error? A loss function declares how costly each mistake is E.g. 0 loss for correct label, 1 loss for wrong label Can weight mistakes differently (e.g. false positives worse than false negatives or Hamming distance over structured labels) We could, in principle, minimize training loss: This is a hard, discontinuous optimization problem Linear Models: Perceptron The perceptron algorithm Iteratively processes the training set, reacting to training errors Can be thought of as trying to drive down training error The (online) perceptron algorithm: Start with zero weights Visit training instances one by one Try to classify If correct, no change! If wrong: adjust weights 9

Examples: Perceptron Separable Case Examples: Perceptron Separable Case 10

Perceptrons and Separability A data set is separable if some parameters classify it perfectly Separable Convergence: if training data separable, perceptron will separate (binary case) Mistake Bound: the maximum number of mistakes (binary case) related to the margin or degree of separability Non-Separable Examples: Perceptron Non-Separable Case 11

Examples: Perceptron Non-Separable Case Issues with Perceptrons Overtraining: test / held-out accuracy usually rises, then falls Overtraining isn t quite as bad as overfitting, but is similar Regularization: if the data isn t separable, weights often thrash around Averaging weight vectors over time can help (averaged perceptron) [Freund & Schapire 99, Collins 02] Mediocre generalization: finds a barely separating solution 12

Problems with Perceptrons Perceptron goal : separate the training data 1. This may be an entire feasible space 2. Or it may be impossible Linear Separators Which of these linear separators is optimal? 13

Objective Functions What do we want from our weights? Depends! So far: minimize (training) errors: This is the zero-one loss Discontinuous, minimizing is NP-complete Not really what we want anyway Maximum entropy and SVMs have other objectives related to zero-one loss Linear Models: Maximum Entropy Maximum entropy (logistic regression) Use the scores as probabilities: Make positive Normalize Maximize the (log) conditional likelihood of training data 14

Derivative for Maximum Entropy Total count of feature n in correct candidates Expected count of feature n in predicted candidates Expected Counts y i P(y x i, w) x i s meal, jail, jail, term, food prison.8.4 The weight for the contextword:jail and cat:prison feature: actual = 1 empirical = 1.2 The optimum parameters are the ones for which each feature s predicted expectation equals its empirical expectation. The optimum distribution is: Always unique (but parameters may not be unique) Always exists (if features counts are from actual data). 15

Maximum Entropy II Motivation for maximum entropy: Connection to maximum entropy principle (sort of) Might want to do a good job of being uncertain on noisy cases in practice, though, posteriors are pretty peaked Regularization (smoothing) Maximum Entropy Separators 16

Maximum Entropy Separators Maximum Entropy 17

Example: NER Smoothing Because of smoothing, the more common prefixes have larger weights even though entire-word features are more specific. State Word Tag Sig Local Context Prev Other at IN x Cur??? Grace NNP Xx Next??? Road NNP Xx Feature Weights Feature Type Feature PERS Previous word at -0.73 Current word Grace 0.03 Beginning bigram <G 0.45 Current POS tag NNP 0.47 Prev and cur tags IN NNP -0.10 Previous state Other -0.70 Current signature Xx 0.80 Prev state, cur sig O-Xx 0.68 Prev-cur-next sig x-xx-xx -0.69 P. state - p-cur sig O-x-Xx -0.20 Total: -0.58 LOC 0.94 0.00-0.04 0.45 0.14-0.92 0.46 0.37 0.37 0.82 2.68 Log-Loss If we view maxent as a minimization problem: This minimizes the log loss on each example One view: log loss is an upper bound on zero-one loss 18

Derivative for Maximum Entropy Big weights are bad Expected count of feature n in predicted candidates Total count of feature n in correct candidates Unconstrained Optimization The maxent objective is an unconstrained optimization problem Basic idea: move uphill from current guess Gradient ascent / descent follows the gradient incrementally At local optimum, derivative vector is zero Will converge if step sizes are small enough, but not efficient All we need is to be able to evaluate the function and its derivative 19

Convexity The maxent objective is nicely behaved: Differentiable (so many ways to optimize) Convex (so no local optima) Convex Non-Convex Convexity guarantees a single, global maximum value because any higher points are greedily reachable Unconstrained Optimization Once we have a function f, we can find a local optimum by iteratively following the gradient For convex functions, a local optimum will be global Basic gradient ascent isn t very efficient, but there are simple enhancements which take into account previous gradients: conjugate gradient, L-BFGs There are special-purpose optimization techniques for maxent, like iterative scaling, but they aren t better 20

Classification Margin (Binary) Distance of x i to separator is its margin, m i Examples closest to the hyperplane are support vectors Marginγof the separator is the minimum m γ m Classification Margin For each example x i and possible mistaken candidate y, we avoid that mistake by a margin m i (y) (with zero-one loss) Margin γ of the entire separator is the minimum m It is also the largest γ for which the following constraints hold 21

Maximum Margin Separable SVMs: find the max-margin w Can stick this into Matlab and (slowly) get an SVM Won t work (well) if non-separable Why Max Margin? Why do this? Various arguments: Solution depends only on the boundary cases, or support vectors (but remember how this diagram is broken!) Solution robust to movement of support vectors Sparse solutions (features not in support vectors get zero weight) Generalization bound arguments Works well in practice for many problems Support vectors 22

Max Margin / Small Norm Reformulation: find the smallest w which separates data Remember this condition? γ scales linearly in w, so if w isn t constrained, we can take any separating w and scale up our margin Instead of fixing the scale of w, we can fix γ = 1 Gamma to w 23

Soft Margin Classification What if the training set is not linearly separable? Slack variablesξ i can be added to allow misclassification of difficult or noisy examples, resulting in a soft margin classifier ξ i ξ i Maximum Margin Non-separable SVMs Add slack to the constraints Make objective pay (linearly) for slack: C is called the capacity of the SVM the smoothing knob Learning: Can still stick this into Matlab if you want Constrained optimization is hard; better methods! We ll come back to this later 24

Maximum Margin Remember SVMs We had a constrained minimization but we can solve for ξ i Giving 25

Hinge Loss Consider the per-instance objective: This is called the hinge loss Unlike maxent / log loss, you stop gaining objective once the true label wins by enough You can start from here and derive the SVM objective Plot really only right in binary case Max vs Soft-Max Margin SVMs: Maxent: You can make this zero but not this one Very similar! Both try to make the true score better than a function of the other scores The SVM tries to beat the augmented runner-up The Maxent classifier tries to beat the soft-max 26

Loss Functions: Comparison Zero-One Loss Hinge Log Separators: Comparison 27