Natural Language Processing with Deep Learning CS224N/Ling284. Richard Socher Lecture 2: Word Vectors

Similar documents
Natural Language Processing with Deep Learning CS224N/Ling284. Christopher Manning Lecture 1: Introduction and Word Vectors

Natural Language Processing

Deep Learning for NLP

11/3/15. Deep Learning for NLP. Deep Learning and its Architectures. What is Deep Learning? Advantages of Deep Learning (Part 1)

Semantics with Dense Vectors. Reference: D. Jurafsky and J. Martin, Speech and Language Processing

Natural Language Processing and Recurrent Neural Networks

Sparse vectors recap. ANLP Lecture 22 Lexical Semantics with Dense Vectors. Before density, another approach to normalisation.

ANLP Lecture 22 Lexical Semantics with Dense Vectors

An overview of word2vec

Deep Learning for NLP Part 2

GloVe: Global Vectors for Word Representation 1

Deep Learning for Natural Language Processing. Sidharth Mudgal April 4, 2017

From perceptrons to word embeddings. Simon Šuster University of Groningen

Deep Learning. Ali Ghodsi. University of Waterloo

CS224n: Natural Language Processing with Deep Learning 1

Lecture 6: Neural Networks for Representing Word Meaning

Natural Language Processing with Deep Learning CS224N/Ling284

Part-of-Speech Tagging + Neural Networks 3: Word Embeddings CS 287

word2vec Parameter Learning Explained

DISTRIBUTIONAL SEMANTICS

Homework 3 COMS 4705 Fall 2017 Prof. Kathleen McKeown

Deep Learning Basics Lecture 10: Neural Language Models. Princeton University COS 495 Instructor: Yingyu Liang

arxiv: v3 [cs.cl] 30 Jan 2016

Algorithms other than SGD. CS6787 Lecture 10 Fall 2017

Neural Networks in Structured Prediction. November 17, 2015

Neural Word Embeddings from Scratch

Online Videos FERPA. Sign waiver or sit on the sides or in the back. Off camera question time before and after lecture. Questions?

Logistic Regression. Will Monroe CS 109. Lecture Notes #22 August 14, 2017

text classification 3: neural networks

Natural Language Processing

Simple Techniques for Improving SGD. CS6787 Lecture 2 Fall 2017

Linear Models in Machine Learning

Seman&cs with Dense Vectors. Dorota Glowacka

Loss Functions and Optimization. Lecture 3-1

Data Mining & Machine Learning

Embeddings Learned By Matrix Factorization

Loss Functions and Optimization. Lecture 3-1

Intelligent Systems (AI-2)

Natural Language Processing with Deep Learning CS224N/Ling284

Algorithms for NLP. Language Modeling III. Taylor Berg-Kirkpatrick CMU Slides: Dan Klein UC Berkeley

CSC321 Lecture 4: Learning a Classifier

Overview Today: From one-layer to multi layer neural networks! Backprop (last bit of heavy math) Different descriptions and viewpoints of backprop

Lecture 7: Word Embeddings

Word Embeddings 2 - Class Discussions

Deep Learning Recurrent Networks 2/28/2018

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann

CS224N: Natural Language Processing with Deep Learning Winter 2018 Midterm Exam

Machine Learning Basics Lecture 3: Perceptron. Princeton University COS 495 Instructor: Yingyu Liang

Language as a Stochastic Process

Distributional Semantics and Word Embeddings. Chase Geigle

CSC321 Lecture 7 Neural language models

Applied Machine Learning Lecture 5: Linear classifiers, continued. Richard Johansson

Lecture 4: Training a Classifier

Midterm sample questions

Naïve Bayes, Maxent and Neural Models

Neural Networks: Backpropagation

Least Mean Squares Regression

CSC321 Lecture 7: Optimization

Deep Learning Sequence to Sequence models: Attention Models. 17 March 2018

Lecture 4: Training a Classifier

Word2Vec Embedding. Embedding. Word Embedding 1.1 BEDORE. Word Embedding. 1.2 Embedding. Word Embedding. Embedding.

15. LECTURE 15. I can calculate the dot product of two vectors and interpret its meaning. I can find the projection of one vector onto another one.

Intelligent Systems (AI-2)

Linear Models for Classification

Lecture 5 Neural models for NLP

Gaussian and Linear Discriminant Analysis; Multiclass Classification

Deep Learning for Natural Language Processing

Page 52. Lecture 3: Inner Product Spaces Dual Spaces, Dirac Notation, and Adjoints Date Revised: 2008/10/03 Date Given: 2008/10/03

Recurrent Neural Networks Deep Learning Lecture 5. Efstratios Gavves

The Kernel Trick, Gram Matrices, and Feature Extraction. CS6787 Lecture 4 Fall 2017

Week 5: Logistic Regression & Neural Networks

Regularization Introduction to Machine Learning. Matt Gormley Lecture 10 Feb. 19, 2018

Bayesian Paragraph Vectors

Comments. Assignment 3 code released. Thought questions 3 due this week. Mini-project: hopefully you have started. implement classification algorithms

Neural Networks for NLP. COMP-599 Nov 30, 2016

Ch.6 Deep Feedforward Networks (2/3)

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Logistic Regression. Professor Ameet Talwalkar. Professor Ameet Talwalkar CS260 Machine Learning Algorithms January 25, / 48

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

ATASS: Word Embeddings

CSC321 Lecture 4: Learning a Classifier

Natural Language Processing with Deep Learning. CS224N/Ling284

Backpropagation Introduction to Machine Learning. Matt Gormley Lecture 12 Feb 23, 2018

Neural Network Training

Statistical NLP for the Web

Topics we covered. Machine Learning. Statistics. Optimization. Systems! Basics of probability Tail bounds Density Estimation Exponential Families

Neural Networks Language Models

Neural Networks. David Rosenberg. July 26, New York University. David Rosenberg (New York University) DS-GA 1003 July 26, / 35

K-means. Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University. November 19 th, Carlos Guestrin 1

Based on slides by Richard Zemel

arxiv: v2 [cs.cl] 1 Jan 2019

Lecture 17: Neural Networks and Deep Learning

CSC321 Lecture 20: Reversible and Autoregressive Models

CS224n: Natural Language Processing with Deep Learning 1

CSC321 Lecture 9: Generalization

Machine Learning and Computational Statistics, Spring 2017 Homework 2: Lasso Regression

15780: GRADUATE AI (SPRING 2018) Homework 3: Deep Learning and Probabilistic Modeling

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

Sequence Modeling with Neural Networks

Transcription:

Natural Language Processing with Deep Learning CS224N/Ling284 Richard Socher Lecture 2: Word Vectors

Organization PSet 1 is released. Coding Session 1/22: (Monday, PA1 due Thursday) Some of the questions from Piazza: sharing the choose-your-own final project with another class seems fine--> Yes* But how about the default final project? Can that also be used as a final project for a different course?--> Yes* Are we allowing students to bring one sheet of notes for the midterm?--> Yes Azure computing resources for Projects/PSet4. Part of milestone 2

Lecture Plan 1. Word meaning (15 mins) 2. Word2vec introduction (20 mins) 3. Word2vec objective function gradients (25 mins) 4. Optimization refresher (10 mins) 3

1. How do we represent the meaning of a word? Definition: meaning (Webster dictionary) the idea that is represented by a word, phrase, etc. the idea that a person wants to express by using words, signs, etc. the idea that is expressed in a work of writing, art, etc. Commonest linguistic way of thinking of meaning: signifier (symbol) signified (idea or thing) 4 = denotation

How do we have usable meaning in a computer? Common solution: Use e.g. WordNet, a resource containing lists of synonym sets and hypernyms ( is a relationships). e.g. synonym sets containing good : e.g. hypernyms of panda : (adj) full, good (adj) estimable, good, honorable, respectable (adj) beneficial, good (adj) good, just, upright (adj) adept, expert, good, practiced, proficient, skillful (adj) dear, good, near (adj) good, right, ripe (adv) well, good (adv) thoroughly, soundly, good (n) good, goodness (n) commodity, trade good, good 5 [Synset('procyonid.n.01'), Synset('carnivore.n.01'), Synset('placental.n.01'), Synset('mammal.n.01'), Synset('vertebrate.n.01'), Synset('chordate.n.01'), Synset('animal.n.01'), Synset('organism.n.01'), Synset('living_thing.n.01'), Synset('whole.n.02'), Synset('object.n.01'), Synset('physical_entity.n.01'), Synset('entity.n.01')]

Problems with resources like WordNet Great as a resource but missing nuance e.g. proficient is listed as a synonym for good. This is only correct in some contexts. Missing new meanings of words e.g. wicked, badass, nifty, wizard, genius, ninja, bombest Impossible to keep up-to-date! Subjective Requires human labor to create and adapt Hard to compute accurate word similarity à 6

Representing words as discrete symbols In traditional NLP, we regard words as discrete symbols: hotel, conference, motel Means one 1, the rest 0s Words can be represented by one-hotvectors: motel = [0 0 0 0 0 0 0 0 0 0 1 0 0 0 0] hotel = [0 0 0 0 0 0 0 1 0 0 0 0 0 0 0] Vector dimension = number of words in vocabulary (e.g. 500,000) 7

Problem with words as discrete symbols Sec. 9.2.2 Example: in web search, if user searches for Seattle motel, we would like to match documents containing Seattle hotel. But: motel = [0 0 0 0 0 0 0 0 0 0 1 0 0 0 0] hotel = [0 0 0 0 0 0 0 1 0 0 0 0 0 0 0] These two vectors are orthogonal. There is no natural notion of similarity for one-hot vectors! Solution: Could rely on WordNet s list of synonyms to get similarity? Instead: learn to encode similarity in the vectors themselves 8

Representing words by their context Core idea: A word s meaning is given by the words that frequently appear close-by You shall know a word by the company it keeps (J. R. Firth 1957: 11) One of the most successful ideas of modern statistical NLP! When a word w appears in a text, its context is the set of words that appear nearby (within a fixed-size window). Use the many contexts of w to build up a representation of w government debt problems turning into banking crises as happened in 2009 saying that Europe needs unified banking regulation to replace the hodgepodge India has just given its banking system a shot in the arm 9 These context words will represent banking

Word vectors We will build a dense vector for each word, chosen so that it is similar to vectors of words that appear in similar contexts. linguistics = 0.286 0.792 0.177 0.107 0.109 0.542 0.349 0.271 Note: word vectors are sometimes called word embeddings or word representations. 10

2. Word2vec: Overview Word2vec (Mikolov et al. 2013) is a framework for learning word vectors. Idea: We have a large corpus of text Every word in a fixed vocabulary is represented by a vector Go through each position t in the text, which has a center word c and context ( outside ) words o Use the similarity of the word vectors for c and o to calculate the probability of o given c (or vice versa) Keep adjusting the word vectors to maximize this probability 11

Word2Vec Overview Example windows and process for computing P w 789 w 7 P w 7=< w 7 P w 78< w 7 P w 7=; w 7 P w 78; w 7 problems turning into banking crises as outside context words in window of size 2 center word at position t outside context words in window of size 2 12

Word2Vec Overview Example windows and process for computing P w 789 w 7 P w 7=< w 7 P w 7=; w 7 P w 78; w 7 P w 78< w 7 problems turning into banking crises as outside context words in window of size 2 center word at position t outside context words in window of size 2 13

Word2vec: objective function For each position t = 1,, T, predict context words within a window of fixed size m, given center word w 9. M L θ = F F P w 789 w 7 ; θ 7N; =IJ9JI 9KL The objective function J θ is the (average) negative log likelihood: 14 Likelihood = θ is all variables to be optimized sometimes called cost or loss function M J θ = 1 T log L(θ) = 1 T S S log P w 789 w 7 ; θ 7N; =IJ9JI 9KL Minimizing objective function Maximizing predictive accuracy

Word2vec: objective function We want to minimize the objective function: M J θ = 1 T S S log P w 789 w 7 ; θ 7N; =IJ9JI 9KL Question: How to calculate P w 789 w 7 ; θ? Answer: We will use two vectors per word w: v U when w is a center word u U when w is a context word Then for a center word c and a context word o: 15 P o c = exp (u Y M v Z ) exp (u U M v Z ) U ]

Word2Vec Overview with Vectors Example windows and process for computing P w 789 w 7 P u^_y`abic v de7y short for P problems into ; u^_y`abic, v de7y, θ P u^_y`abic v de7y P u Z_dcdc v de7y P u 7seder v de7y P u`peqder v de7y problems turning into banking crises as outside context words in window of size 2 center word at position t outside context words in window of size 2 16

Word2Vec Overview with Vectors Example windows and process for computing P w 789 w 7 P u 7s_eder v`peqder P u pc v`peqder P u de7y v`peqder P u Z_dcbc v`peqder problems turning into banking crises as outside context words in window of size 2 center word at position t outside context words in window of size 2 17

Word2vec: prediction function P o c = exp (u Y M v Z ) exp (u U M v Z ) U ] Dot product compares similarity of o and c. Larger dot product = larger probability After taking exponent, normalize over entire vocabulary This is an example of the softmax function R e R e exp (x d ) softmax x d = e = p exp (x 9 ) d 9N; The softmax function maps arbitrary values x d to a probability distribution p d max because amplifies probability of largest x d 18 soft because still assigns some probability to smaller x d Frequently used in Deep Learning

To train the model: Compute all vector gradients! Recall: θ represents all model parameters, in one long vector In our case with d-dimensional vectors and V-many words: Remember: every word has two vectors We then optimize these parameters 19

3. Derivations of gradient Whiteboard see video if you re not in class ;) The basic Lego piece Useful basics: If in doubt: write out with indices Chain rule! If y = f(u) and u = g(x), i.e. y = f(g(x)), then: 20

Chain Rule Chain rule! If y = f(u) and u = g(x), i.e. y = f(g(x)), then: Simple example: 21

Interactive Whiteboard Session! Let s derive gradient for center word together For one example window and one example outside word: 22 You then also need the gradient for context words (it s similar; left for homework). That s all of the parameters θ here.

Calculating all gradients! We went through gradient for each center vector v in a window We also need gradients for outside vectors u Derive at home! Generally in each window we will compute updates for all parameters that are being used in that window. For example: P u 7s_eder v`peqder P u pc v`peqder P u de7y v`peqder P u Z_dcbc v`peqder problems turning into banking crises as outside context words in window of size 2 center word at position t outside context words in window of size 2 23

Word2vec: More details Why two vectors? à Easier optimization. Average both at the end. Two model variants: 1. Skip-grams (SG) Predict context ( outside ) words (position independent) given center word 2. Continuous Bag of Words (CBOW) Predict center word from (bag of) context words This lecture so far: Skip-gram model Additional efficiency in training: 1. Negative sampling So far: Focus on naïve softmax (simpler training method) 24

Gradient Descent We have a cost function J θ we want to minimize Gradient Descent is an algorithm to minimize J θ Idea: for current value of θ, calculate gradient of J θ, then take small step in direction of negative gradient. Repeat. Note: Our objectives are not convex like this :( 25

Intuition For a simple convex function over two parameters. Contour lines show levels of objective function 26

Gradient Descent Update equation (in matrix notation): α = step size or learning rate Update equation (for single parameter): Algorithm: 27

Stochastic Gradient Descent Problem: J θ is a function of all windows in the corpus (potentially billions!) So is very expensive to compute You would wait a very long time before making a single update! Very bad idea for pretty much all neural nets! Solution: Stochastic gradient descent (SGD) Repeatedly sample windows, and update after each one. Algorithm: 28

29

PSet1: The skip-gram model and negative sampling From paper: Distributed Representations of Words and Phrases and their Compositionality (Mikolov et al. 2013) Overall objective function (they maximize): The sigmoid function! (we ll become good friends soon) So we maximize the probability of two words co-occurring in first log à 30

PSet1: The skip-gram model and negative sampling Simpler notation, more similar to class and PSet: We take k negative samples. Maximize probability that real outside word appears, minimize prob. that random words appear around center word P(w)=U(w) 3/4 /Z, the unigram distribution U(w) raised to the 3/4 power (We provide this function in the starter code). The power makes less frequent words be sampled more often 31

PSet1: The continuous bag of words model Main idea for continuous bag of words (CBOW): Predict center word from sum of surrounding word vectors instead of predicting surrounding single words from center word as in skipgram model To make assignment slightly easier: Implementation of the CBOW model is not required (you can do it for a couple of bonus points!), but you do have to do the theory problem on CBOW. 32