Deep learning for Natural Language Processing and Machine Translation

Similar documents
Improved Learning through Augmenting the Loss

High Order LSTM/GRU. Wenjie Luo. January 19, 2016

CSC321 Lecture 10 Training RNNs

a) b) (Natural Language Processing; NLP) (Deep Learning) Bag of words White House RGB [1] IBM

arxiv: v1 [cs.cl] 21 May 2017

From perceptrons to word embeddings. Simon Šuster University of Groningen

Slide credit from Hung-Yi Lee & Richard Socher

Introduction to RNNs!

A QUESTION ANSWERING SYSTEM USING ENCODER-DECODER, SEQUENCE-TO-SEQUENCE, RECURRENT NEURAL NETWORKS. A Project. Presented to

Based on the original slides of Hung-yi Lee

EE-559 Deep learning LSTM and GRU

CSC321 Lecture 15: Recurrent Neural Networks

Recurrent Neural Networks (Part - 2) Sumit Chopra Facebook

Natural Language Processing (CSEP 517): Language Models, Continued

Deep Feedforward Networks. Seung-Hoon Na Chonbuk National University

Deep Learning Sequence to Sequence models: Attention Models. 17 March 2018

Natural Language Processing and Recurrent Neural Networks

arxiv: v3 [cs.lg] 14 Jan 2018

Recurrent Neural Networks. Jian Tang

Deep Learning Recurrent Networks 2/28/2018

Natural Language Processing (CSEP 517): Machine Translation (Continued), Summarization, & Finale

Deep Learning for NLP

Google s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation

arxiv: v1 [cs.cl] 22 Jun 2015

11/3/15. Deep Learning for NLP. Deep Learning and its Architectures. What is Deep Learning? Advantages of Deep Learning (Part 1)

Long Short- Term Memory (LSTM) M1 Yuichiro Sawai Computa;onal Linguis;cs Lab. January 15, Deep Lunch

Deep Learning for NLP Part 2

Part-of-Speech Tagging + Neural Networks 3: Word Embeddings CS 287

Deep Learning and Lexical, Syntactic and Semantic Analysis. Wanxiang Che and Yue Zhang

Lecture 5 Neural models for NLP

Deep Learning. Recurrent Neural Network (RNNs) Ali Ghodsi. October 23, Slides are partially based on Book in preparation, Deep Learning

NEURAL LANGUAGE MODELS

Stephen Scott.

Neural Networks: Backpropagation

Lecture 6: Neural Networks for Representing Word Meaning

Combining Static and Dynamic Information for Clinical Event Prediction

Learning to translate with neural networks. Michael Auli

Conditional Language modeling with attention

arxiv: v1 [stat.ml] 18 Nov 2017

Recurrent Neural Network

Neural Networks Language Models

CS230: Lecture 8 Word2Vec applications + Recurrent Neural Networks with Attention

Lecture 17: Neural Networks and Deep Learning

CSC321 Lecture 16: ResNets and Attention

Multi-Source Neural Translation

TTIC 31230, Fundamentals of Deep Learning David McAllester, April Sequence to Sequence Models and Attention

Gated Recurrent Neural Tensor Network

Artificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino

Long-Short Term Memory and Other Gated RNNs

Deep Learning for Natural Language Processing

arxiv: v2 [cs.cl] 1 Jan 2019

Neural Architectures for Image, Language, and Speech Processing

Deep Learning Tutorial. 李宏毅 Hung-yi Lee

arxiv: v1 [cs.cl] 31 May 2015

Deep Gate Recurrent Neural Network

Sequence Models. Ji Yang. Department of Computing Science, University of Alberta. February 14, 2018

Multi-Source Neural Translation

Deep Neural Machine Translation with Linear Associative Unit

Natural Language Understanding. Recap: probability, language models, and feedforward networks. Lecture 12: Recurrent Neural Networks and LSTMs

Machine Learning for Signal Processing Neural Networks Continue. Instructor: Bhiksha Raj Slides by Najim Dehak 1 Dec 2016

Lecture 15: Exploding and Vanishing Gradients

Self-Attention with Relative Position Representations

Lecture 11 Recurrent Neural Networks I

Recurrent and Recursive Networks

Faster Training of Very Deep Networks Via p-norm Gates

Neural Word Embeddings from Scratch

Coverage Embedding Models for Neural Machine Translation

CKY-based Convolutional Attention for Neural Machine Translation

UNSUPERVISED LEARNING

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Recurrent Neural Networks 2. CS 287 (Based on Yoav Goldberg s notes)

Introduction to Deep Neural Networks

Lecture 11 Recurrent Neural Networks I

Machine Translation. 10: Advanced Neural Machine Translation Architectures. Rico Sennrich. University of Edinburgh. R. Sennrich MT / 26

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Gated Feedback Recurrent Neural Networks

Random Coattention Forest for Question Answering

arxiv: v1 [cs.ne] 19 Dec 2016

Deep Learning for Natural Language Processing. Sidharth Mudgal April 4, 2017

Recurrent Neural Networks. deeplearning.ai. Why sequence models?

arxiv: v2 [cs.ne] 7 Apr 2015

Continuous Space Language Model(NNLM) Liu Rong Intern students of CSLT

Lecture 15: Recurrent Neural Nets

Recurrent Neural Networks Deep Learning Lecture 5. Efstratios Gavves

IMPROVING STOCHASTIC GRADIENT DESCENT

Generating Sequences with Recurrent Neural Networks

Recurrent Neural Networks (RNN) and Long-Short-Term-Memory (LSTM) Yuan YAO HKUST

Neural Network Language Modeling

Inter-document Contextual Language Model

CSC321 Lecture 15: Exploding and Vanishing Gradients

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Recurrent Neural Networks

Sparse vectors recap. ANLP Lecture 22 Lexical Semantics with Dense Vectors. Before density, another approach to normalisation.

ANLP Lecture 22 Lexical Semantics with Dense Vectors

Learning Deep Architectures for AI. Part II - Vijay Chakilam

Review Networks for Caption Generation

An overview of word2vec

CS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes

Key-value Attention Mechanism for Neural Machine Translation

Structured Neural Networks (I)

Transcription:

Deep learning for Natural Language Processing and Machine Translation 2015.10.16 Seung-Hoon Na

Contents Introduction: Neural network, deep learning Deep learning for Natural language processing Neural network for classification Word embedding General architecture for sequential labeling Recent advances Neural machine translation Future plan 2

Deep learning: Motivation 기계학습방법에서자질추출 (Feature extraction) Handcrafted features 사용 자질추출단계가자동화되지는않음 지속적으로자질개선필요 성능개선및튜닝요구 Feature vector f(x; w) = argmax y w Ψ(x, y) Deep learning 자질추출단계의제거또는간소화 정교한자질을비선형학습과정에내재시킴

Multi-layer perceptron Deep learning 상위은닉층은하위은닉층의출력에대한추상화 비선형성모델 다층 NN 구조로추상자질내재가능 자질튜닝절차를단순화시킴 Abstraction 일반적인기계학습 Raw data Hand-crafted Features Trainable Classifier 딥러닝 Raw data Trainable Features Trainiable Classifier

A neuron Neural Network A general computational unit that takes n inputs and produces a single output Sigmoid & binary logistic regression unit The most popular neurons Takes an n-dimensional input vector x and produces the scalar activation (output) a x 1 w 1 a = 1 1 + exp(w T x + b) x 2 w 2 σ a x n w n b 5

Neural network: Setting for Classification Input layer: Feature values Output layer: Scores of labels Softmax layer: Normalization of output values To get probabilities of output labels given input K: the number of labels Softmax layer y i = exp(y i) exp(y t ) y 1 y 2 y K 1 y K y 3 Output layer Hidden layer Input layer 6

Neural network: Training Training data: Tr = (x 1, g 1 ),, (x N, g N ) x i : i-th input feature vector g i {1,, K}: i-th target label Objective function Negative Log-likelihood (NLL) L = x,g T log P(g x) 7

Neural network: Training Stochastic gradient method 1. Randomly sample (x, g) from training data 2. Define NLL for (x, g) L = log P(g x) for each weight matrix W θ 3. Compute gradients : L W 4. Update weight matrix W: W W η L W Iterate the above procedure 8

Neural network: Backpropagation Output layer L = log P g x = log exp(y g) exp(y i ) = y g log exp(y i ) softmax y 1 y 2 y i y K 1 y K Compute delta for i-th output node δ i o = L y i = δ i, g exp y i exp y j = δ i, g P(i x) Vector form: δ o = 1 g P(1 x) P(K x) 9

Neural network: Backpropagation Output weight matrix W softmax y g log exp(y i ) y 1 y 2 y i y K 1 y K δ i o Output layer w ij h 1 h j h m Hidden layer z 1 z j z m Compute gradient of w ij h j = g z j L w ij = L y i y i w ij = δ i o h j L W = δo h T 10

Neural network: Backpropagation Hidden layer y g log exp(y i ) softmax y 1 y 2 y i y K 1 y K δ i o Output layer w ij h 1 z 1 Compute delta for j-th hidden node δ j h = L z j = L h j h j z j = h j z j i L y i y i h j = g z j δ h = g z W T δ o h j z j h m z m Hidden layer i δ i O w ij h j = g z j 11

Neural network: Backpropagation Hidden weight matrix U y g log exp(y i ) h 1 z 1 h j h δ z j j u jk h m z m Hidden layer h j = g z j x k Compute gradient of u ij L u jk = L z j z j u jk = δ j h x k L U = δh x T 12

Deep learning for NLP NN for Word embedding Raw corpus Unsupervised Learning word embedding matrix Applicationspecific NN Annotated corpus Supervised Initialize lookup table Application-specific neural network Lookup table is further fine-tuned 13

Word embedding: Distributed representation Distributed representation n-dimensional latent vector for a word Semantically similar words are closely located in vector space 14

Word 0 1 0 One-hot vector e ( V -dimensional vector) Word embedding matrix: Lookup Table V L= the cat mat x =L e L = R d V d Word vector x is obtained from one-hot vector e by referring to lookup table 15

Word embedding matrix: context input layer word seq w 1 w n input layer n d dim input vector concatenation d-diml vec d-dim vec d-dim vec Lookup table Input layer L e wt n+1 L e wt 2 L e wt 1 w t n+1 w t 2 w t 1 n context words e w : one hot vector w 16

Neural Probabilistic Language Model (Bengio 03) Language models The probability of a sequence of words N-gram model P(w i w 1,, w i 1 ) P(w i w i n 1,, w i 1 ) Neural probabilistic language model Estimate P(w i w i n 1,, w i 1 ) by neural networks for classification x: Concatenated input features (input layer) y = U tanh(d + Hx) y = Wx + b + U tanh(d + Hx) 17

Neural probabilistic language model P w t w t n+1,, w t 1 = exp(y w t ) exp(y t ) w t Softmax layer y 1 y 2 y V Output layer W U H Hidden layer Lookup table Input layer w t n+1 w t 2 w t 1 18

Neural Probabilistic Language Model P w t w t n+1,, w t 1 = Directly connected to output layer Softmax= exp(y w t ) exp(y t ) tanh Normalization for probabilistic value Output layer Each node indicates a specific word Hidden Layer Input Lookup table layer w t n+1 w t 2 w t 1 19

NPLM: Discussion Limitation: Computational complexity Softmax layer requires computing scores over all vocabulary words Vocabulary size is very large 20

Ranking Approach for Word Embedding Idea: Sampling a negative example (Collobert & Weston 08) s: a given sequence of words (in training data) s : a negative example the last word is replaced with another word f(s): score of the sequence s Goal: makes the score difference (f(s) f(s )) large Various loss functions are possible Hinge loss: max(0, 1 f s + f s ) 21

Ranking Approach for Word Embedding (Collobert and Weston 08) f(w t n+1,, w t 1, w t ) wtanh Hx + b + d w f(w t n+1,, w t 1, w t ) wtanh Hx + b + d w Hidden layer Lookup table H x w t n+1 w t 2 w t 1 Positive example H x w t w t n+1 w t 2 w t 1 w t Negative example 22 Input layer Sampled Last word

Recurrent Neural Network (RNN) based Language Model NPLM: only (n-1) previous words are conditioned P(w i w i n 1,, w i 1 ) RNNLM: all previous words are conditioned P(w i w 1,, w i 1 ) Recurrent neural network 23

Recurrent Neural Network y t 1 y t y t+1 h t = f(x t, h t 1 ) U h t 1 W h t h t+1 z t = Wh t 1 + Vx t h t = g(z t ) V y t = Uh t x t 1 x t x t+1 24

Recurrent Neural Network: Backpropagation Through Time Objective function: L = t L t y t U y t+1 y T U U L U = t δ t o h t T W h t W h t+1 W h T L W = t δ h T t h t 1 V V V V L V = t δ t h x t T x t x t+1 x T 1 x T δ h t = g z t+1 W T h δ t+1 + U T δ t o 25

Recurrent Neural Network: BPTT Objective function: L = t L t y t 1 h t 1 W y t h t U W y t+1 h t+1 Compute gradient of L(t) w.r.t. params L(t) U = δ t o h t T h δ t 1,t h δ t 2,t δ h t,t = U T δ t o h = g z t W T δ t,t = g z t 1 W T δ t 1,t Given specific time t h x t 1 x t V V x t+1 L(t) W = δ t,t h T h t 1 L(t) V L(t) C(w) = h T + δ t 1,t h t 2 + + δ h T 1,t h 0 = δ t,t h x T t + + δ h T 1,t x 0 t t:w t =w x δ t,t Lookup table

Recurrent Neural Network: BPTT Vanishing gradient & gradient explosion problems L(t) W = δ t,t h T h t 1 h T + δ t 1,t h t 2 + + δ h T 1,t h 0 h δ t 1,t = g z t W T h δ t,t h δ t 2,t = g z t 1 W T h δ t 1,t = g z t 1 g z t W T W T δ t,t h g z k g z t 1 g z t W T W T W T Can easily become a very small or large number 27

Recurrent Neural Network: BPTT Solutions to the exploding and vanishing gradients 1. Instead of initializing W randomly, start off from an identify matrix initialization 2. Use the Rectified linear units (ReLU) instead of the sigmoid function The derivative for the ReLU is either 0 or 1 28

Experiments: NPLM NPLM 29

Experiments: RNN LM 30

Long Short Term Memory (LSTM) RNN: Very hard to actually train long-term dependencies exploding & vanishing gradients LSTM: makes it easier for RNNs to capture long-term dependencies Using gated units Traditional LSTM (Hochreiter and Schmidhuer, 98) Introduces input gate & output gate Limitation: The output is close to zero as long as the output gate is closed. Modern LSTM: Uses forget gate (Gers et al 00) Variants of LSTM Add peephole connections (Gers et al 02) Allow all gates to inspect the current cell state even when the output gate is closed. 31

Long Short Term Memory (LSTM) h (t) = f(x t, h (t 1) ) i t = σ W i x t + U i h t 1 (Input gate) f t = σ W f x t + U f h t 1 (Forget gate) o t = σ W o x t + U o h t 1 (Output/Exposure gate) c t = tanh W c x t + U c h t 1 (New memory cell) c (t) = f (t) c t 1 + i (t) c t h (t) = o (t) tanh(c t ) (Final memory cell) 32

Long Short Term Memory (LSTM) 33

LSTM: Memory cell 34

Long Short Term Memory (LSTM) Input gate: Whether or not the input is worth preserving and thus is used to gate the new memory Forget gate: Whether the past memory call is useful for the computation of the current memory cell Final memory generation: Takes the advices of the forget and input gates to produce the final memory Output/Exposure gate: What parts of the memory needs to be explored in the hidden state The purpose is to separate the final memory from the hidden state. The final memory contains a lot of information that is not necessarily required to be saved in the hidden state 35

Gated Recurrent Units (Cho et al 14) h (t) = f(x t, h (t 1) ) Alternative architecture to handle long-term dependencies z t = σ W z x t + U z h t 1 (Update gate) r t = σ W r x t + U r h t 1 (Reset gate) h t = tanh r (t) Uh t 1 + Wx (t) (New memory) h (t) = (1 z (t) ) h t + z (t) h t 1 (Hidden state) 36

Gated Recurrent Units 37

Gated Recurrent Units New memory generation: A new memory h (t) is the consolidation of a new input word x (t) with the past hidden state h (t 1) Reset gate: Determining how important h (t 1) is to the summarization h (t). It can completely diminish past hidden state if it is irrelevant to the computation of the new memory Update gate: Determining how much of h (t 1) should be carried forward to the next stage Hidden state: The hidden state h (t) is generated using the past hidden input h (t 1) and the new memory generated h (t) with the advice of the update gate 38

LSTM vs. GRU 39

Neural Machine Translation: RNN Encoder-Decoder (Cho et al 14) Computing the log of translation probability Log P(y x) by two RNNs 40

Neural Machine Translation RNN Encoder Decoder RNN Search RNN Encoder Decoder with Attention RNN Search Large Vocabulary Sequence-to-Sequence Model LSTM Rare word model 41

Neural Machine Translation: RNN Encoder-Decoder (Cho et al 14) RNN for the encoder: input x hidden state c h t = f(h t 1, x t ) Encode a variable-length sequence into a fixed-length vector representation Reads each symbol of an input sequence x sequentially After reading the end of the sequence, we obtain a summary c of the whole input sequence: c = h T 42

RNN Encoder-Decoder (Cho et al 14) RNN for the decoder: Hidden state c Output y h t = f h t 1, y t 1, c P(y t y t 1,, y 1, c) = g h t, y t 1, c Decode a given fixed-length vector representation back into a variable-length sequence Generate the output sequence by predicting the next symbol y t given the previous hidden state Both y t and h t are conditioned on y t 1 How to define f? Gated recurrent unit (GRU) is used to capture long-term dependencies 43

Using RNN Encoder-Decoder for Statistical Machine Translation Statistical machine translation Generative model: P f e p e f P f p e f : Translation model, p(f): Language model Log-linear model: logp f e = w n f n (f, e) + logz e In Phrase-based SMT logp(e f) is factorized into the translation probabilities of matching phrases in the source and target sents Scoring phrase pairs with RNN Encoder-Decoder Train the RNN Encoder-Decoder on a table of phrase pairs Use its scores as additional features Ignore the normalized frequencies of each phrase pair Existing translation prob in the table already reflects the frequencies of the phrase pairs 44

RNN Encoder-Decoder: Experiments Training # hidden units: 1,000, # dim for word embedding: 100 Adadelta & stochastic gradient descent Use Controlled vocabulary setting Tmost frequent 15,000 words for both source & target languages All the out-of-vocabulary words are mapped to UNK English-to-French translation 45

Neural Machine Translation: Sequence-to-sequence (Sutskever et al 14) <EOS>: the end-of-sentence token 1) a LSTM reads input sentence x which generates the last hidden state v 2) a standard LSTM-LM computes the probability of y 1,, y T given the initial hidden state v P y 1,, y T x 1,, x T = t P(y t v, y 1,, y t 1 ) P(y t v, y 1,, y t 1 ): Taking Softmax over all words 46

Can the LSTM reconstruct the input sentence? Can this scheme learn the identity function? Answer: it can, and it can do it very easily. It just does it effortlessly and perfectly. Slide from http://www.cs.tau.ac.il/~wolf/deeplearningmeeting/pdfs/sup+lstm.pdf 47

Sequence-to-Sequence Model (Sutskever et al 14) Two different LSTMs one for the input sentence Another for the output sentence Deep LSTM (with 4 layer) 1,000 cell at each layer, 1,000 dimensional word embedding Deep LSTM uses 8000 numbers for sentence representation Revising source sentence, C, B, A A, B, C, Backprop can notice the short-term dependencies first, and slowly extend them to long range dependencies Ensemble of deep LSTMs LSTMs trained from Different initialization 48

Sequence-to-Sequence Model : End-to-end Translation Experiments English-to-French translation No SMT is used for reference model Input/Output vocabulary size: 160,000/80,000 Training objective: maximizing (S,T) logp(t S) Decoding: Beam search 49

Sequence-to-Sequence Model : Re-scoring Experiments Rescoring the 1000-best list of the baseline system 50

Sequence-to-Sequence Model : Learned representation 2-dimensional PCA projection of the LSTM hidden states 51

Sequence-to-Sequence Model: Learned representation 2-dimensional PCA projection of the LSTM hidden states 52

Sequence-to-Sequence Learning: Performance vs Sentence Length 53

Sequence-to-Sequence Learning: Performance on rare words 54

RNN Search: Jointly Learning to Align and Translate (Bahdanau et al 15) RNN Search: Extended architecture of RNN Encoder-Decoder Encodes the input sentence into a seq of vectors using a bidirectional RNN Context vector c Previously, the last hidden state (Cho et al 14) Now, mixture of hidden states of input sentence at generating each target word During decoding, chooses a subset of these vectors adaptively 55

RNN Search: Components p y i y 1,, y i 1 = g y i 1, s i, c i s i = f(s i 1, y i 1, c i ) An RNN hidden state c i = j α ij h j α ij = exp(e ij) k exp(e ik ) e ij = a(s i 1, h j ) Alignment model 56

RNN Search: Encoder Bidirectional RNN 57

RNN Search: Decoder compute alignment compute context generate new output compute new decoder state 58

RNN Search: Alignment model e ij = v T tanh Ws i 1 + Vh j Directly computes a soft alignment α ij = exp(e ij) k exp(e ik ) Expected annotation Alignment model is jointly trained with all other components 59

RNN Search: Experiment English-to-French Model RNN Search, 1,000 units Baseline RNN Encoder-Decoder, 1000 units Data English to French translation, 348 million words 30000 words + UNK token for the networks, all words for Moses Training Stochastic gradient descent (SGD) with a minibatch of 80 Decoding Beam search (Sutskever 14) 60

RNN Search: Experiment English-to-French 61

RNN Search: Large Target Vocabulary (Jean et al 15) Decoding in RNN search P y t y <t, x = 1 Z exp w t T φ y t 1, z t, c t + b t Normalization constant Z = k:yk V exp w t T φ y t 1, z t, c t + b t Computationally inefficient Idea: Use only small V of target vocabulary to approximate Z Based on importance sampling (Bengio and Senecal, 08) 62

RNN Search: Large Target Vocabulary (Jean et al 15) logp y t y <t, x = E y t k:y k V P(y k y <t, x) E y k E y j = w j T φ y j 1, z j, c j + b j Energy function Approximation by importance sampling k:y k V ω k k :y k V ω k E y k E P E y Expected gradient of the energy ω k = exp{e y k logq(y k )} Proposal distribution 63

RNN Search LV: Experiments Setting (Vocabulary size) RNN Search: 30,000 for En-Fr, 50,000 for En-Ge RNN Search LV: 500,000 source and target words 64

Sequence-to-Sequence Model: Rare Word Models (Luong et al 15) Extend LSTM for NMT (Suskever 14) Translation-specific approach For each OOV word, we make a pointer to its corresponding word in the source sentence The pointer information is later utilized in a postprocessing step Directly translates OOV using a original dictionary or with the identify translation, if no translation is found 65

Rare Word Models (Luong et al 15) Replacing rare words with UNK Positional information is stored in target sentence 66

Rare Word Models: Experiments 67

Grammar as Translation (Vinyals et al 15) Extended Sequence-to-Sequence Model Two LSTMs + Attention (Bahdanau et al 15) Linearizing parse trees: Depth-first traversal 68

Grammar as Translation: Linearizing Parse Trees 69

Sequence-to-Sequence Model: Parsing - Experiment Results 70

Neural Network Transition-Based Parsing (Weiss et al 15) Structured learning! 71

NN Transition-Based Parsing: Experiments 72

Neural Network Graph-Based Parsing (Pei et al 15) 73

Neural Network Graph-Based Parsing (Pei et al 15) 74

NN Graph-Based Parsing: Experiments 75

Discussion: Research Topics Neural Machine Translation DNNs represent words, phrases, and sentences in continuous space How to utilize more syntactic knowledge? Recursive recurrent NN (Liu et al 14) Deep architecture may approximately represents input structure Neural Dependency Parsing Neural Language Model 76

Reference Mikolov Tomá¹, Karafiát Martin, Burget Luká¹, Èernocký Jan, Khudanpur Sanjeev: Recurrent neural network based language model, In: Proceedings of the 11th Annual Conference of the International Speech Communication Association (INTERSPEECH 2010) Yoshua Bengio, Jerome Louradour, Ronan Collobert and Jason Weston, Curriculum Learning, ICML '09 Yoshua Bengio, Réjean Ducharme and Pascal Vincent, A Neural Probabilistic Language Model, NIPS '01 Kyunghyun Cho, Bart van Merrienboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, Yoshua Bengio, Learning Phrase Representations using RNN Encoder Decoder for Statistical Machine Translation, 2014 S. Hochreiter, J. Schmidhuer, "Long short-term memory," Neural Computation, vol.9, no.8, pp.1735-1780, 1997. F. A. Gers, N. N. Schraudolph, J. Schmidhuer, Learning Precise Timing with LSTM Recurrent Networks, JMLR, 2002 Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, Yoshua Bengio, Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling, 2014 Ilya Sutskever, Oriol Vinyals, Quoc V. Le, Sequence to Sequence Learning with Neural Networks, NIPS '14 N. Kalchbrenner and Phil Blunsom, Recurrent continuous translation models, EMNLP 13 Dzmitry Bahdanau, KyungHyun Cho, Yoshua Bengio, Neural Machine Translation by Jointly Learning to Align and Translate, ICLR 15 Minh-Thang Luong, Ilya Sutskever, Quoc V. Le, Oriol Vinyals, Wojciech Zaremba, Addressing the Rare Word Problem in Neural Machine Translation, ALC 15 S ebastien Jean, Kyunghyun Cho, Roland Memisevic, and Yoshua Bengio. 2015. On using very large target vocabulary for neural machine translation. In ACL. Oriol Vinyals, Lukasz Kaiser, Terry Koo, Slav Petrov, Ilya Sutskever, Geoffrey Hinton, Grammar as a Foreign Language, 15 David Weiss, Chris Alberti, Michael Collins, Slav Petrov, Structured Training for Neural Network Transition-Based Parsing, ACL '15 Wenzhe Pei, Tao Ge, Baobao Chang, An Effective Neural Network Model for Graph-based Dependency Parsing, EMNLP '15 Shujie Liu, Nan Yang, Mu Li, and Ming Zhou, A Recursive Recurrent Neural Network for Statistical Machine Translation, ACL '14 Jiajun Zhang and Chengqing Zong, Deep Neural Networks in Machine Translation: An Overview, IEEE Intelligent Systems 15 77