Dependency grammar. Recurrent neural networks. Transition-based neural parsing. Word representations. Informs Models

Size: px
Start display at page:

Download "Dependency grammar. Recurrent neural networks. Transition-based neural parsing. Word representations. Informs Models"

Transcription

1

2 Dependency grammar Morphology Word order Transition-based neural parsing Word representations Recurrent neural networks Informs Models

3 Dependency grammar Morphology Word order Transition-based neural parsing Word representations Recurrent neural networks

4

5 Dependency Grammar I Modern theories of dependency grammar originate with Lucien Tesnière I Reference: Lucien Tesnière (1959). Éléments de syntaxe structurale, Klincksieck, Paris. ISBN I Underlying ideas date back to Panini and his system of karakas I Di erent contemporary frameworks of dependency grammar, including the Prague School s Functional Generative Description, Melcuk s Meaning-Text Theory, and Hudson s Word Grammar.

6 Dependency Grammar The sentence is an organized whole, whose constituent elements are words. [1.2] Every word that belongs to a sentence ceases by itself to be isolated as in the dictionary. Between the word and its neighbors, the mind perceives connections, the totality of which forms the structure of the sentence. [1.3] The structural connections establish dependency relations between the words. Each connection in principle unites a superior term and an inferior term. [2.1] The superior term receives the name governor. Theinferiortermreceivesthenamesubordinate. Thus, in the sentence Alfred parle [... ], parle is the governor and Alfred the subordinate. from: Tesnière (1959)

7 Advantages of Dependency Grammars I a completely word-based framework (no phrasal projections). I most dependency grammar frameworks are non-derivational and mono-stratal. I allows for a surface-level syntactic account of languages with flexible word order and syntactic constructions with discontinuous elements. However, these syntactic phenomena raise also challenging questions about the dependency grammar formalism and the notion of projectivity of dependency structures.

8 Parsing with Dependency Grammars I Parsing a sentence is not a goal in itself, but ultimately needs to help provide an adequate answer to the question: Who did what to whom, when, where, and why? In other words: syntactic structure needs to be linked in a systematic fashion to semantic representation/interpretation. I Dependency grammar o ers a direct interface between syntax and semantics: dependency relations between a governor (lexical head) and its lexical dependents can link lexical representations of the main participants of an event or state of a airs with lexical representations of the cicrumstances under which they occurred or hold. I Parsing with dependency grammars benefits from the lexicalist character of dependency relations. This is beneficial, inter alia, for parsing long-distance dependencies and coordinations (see Kübler and Prokic 2006)

9 UD English treebank: treatment of nominal arguments nsubj aux root obj punct det compound you should get a cocker spaniel. you should get a cocker spaniel. PRON AUX VERB DET NOUN NOUN PUNCT

10 UD English treebank: treatment of PP adjuncts root obl punct nsubj obj case He announced this in January : he announce this in January : PRON VERB PRON ADP PROPN PUNCT

11 UD English treebank: treatment of clausal subjects punct root csubj mark obj obl case Great to have you on board! great to have you on board! ADJ PART VERB PRON ADP NOUN PUNCT

12 UD English treebank: treatment of relative clauses punct nsubj root obj advmod acl:relcl det det nsubj amod Every move Google makes brings this particular future closer. every move Google make bring this particular future closer. DET NOUN PROPN VERB VERB DET ADJ NOUN ADV PUNCT

13 UD English treebank: treatment of relative clauses vocative punct nsubj acl:relcl nsubj root obj punct Malach, What you say makes sense. Malach, what you say make sense. PROPN PUNCT PRON PRON VERB VERB NOUN PUNCT

14 UD English treebank: treatment of direct questions advmod auxpass nsubjpass advmod root nmod nmod:tmod punct Why were they suddenly acted on Saturday? why be they suddenly act on Saturday? ADV AUX PRON ADV VERB ADP PROPN PUNCT

15 Heads and Dependents Tests for identifying a head H and a dependent D in a syntactic construction C: 1. H determines the syntactic category of C and can often replace C. 2. H determines the semantic category of C; D gives semantic specification 3. H is obligatory; D may be optional. 4. H selects D and determines whether D is obligatory or optional. 5. The form of D depends on H (agreement or government). 6. The linear position of D is specified with reference to H. from: Kübler et. al. (2009), p.3f.

16 Heads and Dependents: Some Unclear Cases I Auxiliary-main-verb constructions I determiner-adjective-noun constructions I prepositional phrases I Coordination structures The answer often depends on di erent purposes that the dependency structure is put to use for.

17 Case Study: Strong and Weak Adjectives in Dutch (1) a. de bruine beer the brown [weak] beer [masc] b. een a the brown beer bruine brown a brown beer beer [strong] beer [masc] c. de bruine beest the brown [weak] animal [neut] d. een a the brown animal bruin beest brown [strong] animal [neut] a brown animal

18 Universal Dependency Initiative I objective: develop cross-linguistically consistent treebank annotation for many languages I goal: facilitate multilingual parser development, cross-lingual learning, and parsing research from a language typology perspective I strategy: provide a universal inventory of categories and guidelines to facilitate consistent annotation of similar constructions across languages, while allowing language-specific extensions when necessary

19 UD annotations across languages some examples

20 Universal Dependency Relations

21 Universal Tagset Open class words Closed class words Other ADJ ADP PUNCT ADV AUX SYM INTJ CCONJ X NOUN DET PROPN VERB NUM PART PRON SCONJ

22 Some Definitions: Sentence and Arc Labels Definition 2.1. A sentence is a sequence of tokens denoted by: S = w 0 w 1...w n,wherew 0 = root Definition 2.2. Let R = {r 1,...,r m } be a finite set of possible dependency relation types that can hold between any two words in a sentence. A relation type r 2 R is additionally called an arc label. Acknowledgement: Definitions ; , and Notation and are all taken from Kübler, McDonald, and Nivre (2009), chapt. 2

23 Dependency Structures and Dependency Trees Definition 2.3. A dependency graph G =(V, A) isalabeled directed graph (digraph) in the standard graph-theoretic sense and consists of nodes, V,andarcs,A, suchthatforsentence S = w 0 w 1...w n and label set R the following holds: 1. V {w 0, w 1,...,w n } 2. A V R V 3. if (w i, r, w j ) 2 A then (w i, r 0, w j ) /2 A for all r 0 6= r The spanning node set V S = {w 0, w 0,...,w n } contains all and only the words of a sentence, including w 0 = root.

24 Dependency Trees Definition 2.4. A well-formed dependency graph G =(V, A) for an input sentence S and dependency relation set R is any dependency graph that is adirectedtreeoriginating out of node w 0 and has the spanning node set V = V S.Wecallsuch dependency graphs dependency trees.

25 Unique Head Property Remark: Dependency trees rule out the following dependency configuration: arc 1 arc 2 head dep head Some putative counterexample: In cases of VP coordination, as in Sandy listened and smiled, itappearsatleastpausibleto establish a dependency relation between each verbal head to the nominal dependent.

26 Some Notation Notation 2.6. The notation w i! w j indicates the unlabeled dependency relation (or dependency relation for short) in a tree G =(V, A). That is, w i! w j if and only if (w i, r, w j ) 2 A for some r 2 R. Notation 2.7. The notation w i! w j indicates the reflexive transitive closure of the dependency relation in a tree G =(V, A). That is, w i! w j if and only if i = j (reflexive) or both w i! w i 0 and w i 0! w j hold (for some w i 0 2 V ). Notation 2.8. The notation w i $ w j indicates the undirected dependency relation in a tree G =(V, A). That is, w i $ w j if and only if either w i! w j or w j! w i. Notation 2.9. The notation w i $ w j indicates the reflexive transitive closure of the undirected dependency relation in a tree G =(V, A). That is, w i $ w j if and only if i = j (reflexive) or both w i $ w i 0 and w i 0 $ w j hold (for some w i 0 2 V ).

27 Connectedness AdependencytreeG =(V, A) satisfiestheconnectedness property, whichstatesthatforallw i, w j 2 V it is the case that w i $ w j. That is, there is a path connecting every two words in a dependency tree when the direction of the arc (dependency relation) is ignored.

28 (Non-)Projective Dependency Trees Definition An arc (w i, r, w j ) 2 A in a dependency tree G =(V, A) isprojective if and only if w i! w k for all i < k < j when i < j, orj < k < i when j < i. Definition AdependencytreeG =(V, A) isaprojective dependency tree if (1) it is a dependency tree (definition 2.4), and (2) all (w i, r; w j ) 2 A are projective. Definition A dependency tree G = (V;A) is a non-projective dependency tree if (1) it is a dependency tree (definition 2.4), and (2) it is not projective.

29 Converting Non-projective to Projective Dependency Trees PU root ATT TMP PC DET SBJ VC ATT A hearing is scheduled on the issue today. VC:TMP PU root SBJ:ATT PC DET SBJ VC ATT A hearing is scheduled on the issue today.

30 Adependencygammartreebank A dependency gammar treebank consists of pairs of sentences S and their corresponding dependency trees G : T = {(S d, G d )} T d=0 The dependency trees G can be obtained by I manual annotation by one or more human annotators I automatically annotated by a parser I derived automatically by a conversion algorithm from a constituent grammar treebank

31 Tübingen Treebank of Written German (TüBa-D/Z) I developed by my research group at the Seminar für Sprachwissenschaft at the University of Tübingen since I language data taken from the German newspaper die tageszeitung (taz). I largest manually annotated treebank for German I total of 104,787 sentences I average sentence length: 18.7 words per sentence. I total number of tokens: 1,959,474.

32 Tübingen Treebank of Written German (TüBa-D/Z) I orginally annotated for constituent structure I now also available in dependency structure format I The annotation guidelines are published in the Stylebook for the Tübingen Treebank of Written German (TüBa-D/Z) upload/ascl/\tuebadz-stylebook-1508.pdf I Information on how to obtain the data can be found at: resources/corpora/tueba-dz.html

33

34 =<, > =

35 Θ()

36

37

38 = =< σ,β, > σ σ : β β : (,, )

39 = =< σ,β, > σ σ : β β : (,, )

40 σ β

41 σ =[, ] β =[, ] = {(,, )}

42 0 () ([ 0 ], [ 1... ], ) ([ 0 ], [], ) 0 =

43

44

45

46 [σ ] σ [ β] β

47 (σ, [ β], ) ([σ ],β,) ([σ ],β,) ([σ ]β, (,, )) 0 ([σ ],β,) ([σ ],β, (,, ))

48 0, =( 0, 1,..., ) 0 0 () 1... = ( 1 ) =<, >

49 σ β...

50 σ β...

51 σ β...

52 σ β... 1 = {(,, )}

53 σ β... 1 = {(,, )} 1

54 σ β... 1 = {(,, )} 1 1

55 σ β... 1 = {(,, )} = 1 {(,, )}

56 σ β... 1 = {(,, )} = 1 {(,, )} 3 = 2 {(,, )}

57 σ β... 1 = {(,, )} = 1 {(,, )} 3 = 2 {(,, )} 4 = 3 {(,, )}

58

59

60

61 ,

62

63 () ( {}) ()

64 () ( {}) () ( )= {

65 (σ, [ β], ) ([σ ],β,) ([σ ],β,) ([σ ]β, (,, )) 0 ([σ ],β,) ([σ ],β, (,, )) ([σ ],β,) ([σ ],β,)

66

67 σ β...

68 σ β...

69 σ β...

70 σ β... 1 = {(,, )}

71 σ β... 1 = {(,, )} 2 = 1 {(,, )}

72 σ β... 1 = {(,, )} 2 = 1 {(,, )} [σ 1 2 ] 2

73 σ β...

74 σ β...

75 σ β...

76 σ β... 1 = {(,, )}

77 σ β... 1 = {(,, )} 1

78 σ β... 1 = {(,, )} 1 1

79 σ β... 1 = {(,, )} 1 1 1

80 σ β... 1 = {(,, )} = 1 {(,, )}

81 σ β... 1 = {(,, )} = 1 {(,, )} 3 = 2 {(,, )}

82 σ β... 1 = {(,, )} = 1 {(,, )} 3 = 2 {(,, )} 4 = 3 {(,, )}

83 σ β... 1 = {(,, )} = 1 {(,, )} 3 = 2 {(,, )} 4 = 3 {(,, )} 5 = 4 {(,, )}

84 (σ, [ β], ) ([σ ],β,) ([σ ], [ β], ) ([σ ],β, (,, )) ([σ ], [ β], ) (σ, [ β], (,, )) 0 (,, ) ([σ ],β,) (σ, β, ) (,, )

85 (σ, [ β], ) ([σ ],β,) ([σ ], [ β], ) ([σ ],β, (,, )) ([σ ], [ β], ) (σ, [ β], (,, )) 0 (,, ) ([σ ],β,) (σ, β, ) (,, )

86 σ β...

87 σ β...

88 σ β... 1 = {(,, )}

89 σ β... 1 = {(,, )} 2 = 1 {(,, )}

90 σ β... 1 = {(,, )} 2 = 1 {(,, )} 3 = 2 {(,, )}

91 σ β... 1 = {(,, )} 2 = 1 {(,, )} 3 = 2 {(,, )} 4 = 3 {(,, )}

92 σ β... 1 = {(,, )} 2 = 1 {(,, )} 3 = 2 {(,, )} 4 = 3 {(,, )} 5 = 4 {(,, )}

93 σ β... 1 = {(,, )} 2 = 1 {(,, )} 3 = 2 {(,, )} 4 = 3 {(,, )} 5 = 4 {(,, )} 4 5

94

95 2 :

96

97

98

99

100 0 () () () ()

101 O()

102 O()

103 O() 1

104 O() 1 1

105

106

107

Computational Linguistics

Computational Linguistics Computational Linguistics Dependency-based Parsing Clayton Greenberg Stefan Thater FR 4.7 Allgemeine Linguistik (Computerlinguistik) Universität des Saarlandes Summer 2016 Acknowledgements These slides

More information

Computational Linguistics. Acknowledgements. Phrase-Structure Trees. Dependency-based Parsing

Computational Linguistics. Acknowledgements. Phrase-Structure Trees. Dependency-based Parsing Computational Linguistics Dependency-based Parsing Dietrich Klakow & Stefan Thater FR 4.7 Allgemeine Linguistik (Computerlinguistik) Universität des Saarlandes Summer 2013 Acknowledgements These slides

More information

S NP VP 0.9 S VP 0.1 VP V NP 0.5 VP V 0.1 VP V PP 0.1 NP NP NP 0.1 NP NP PP 0.2 NP N 0.7 PP P NP 1.0 VP NP PP 1.0. N people 0.

S NP VP 0.9 S VP 0.1 VP V NP 0.5 VP V 0.1 VP V PP 0.1 NP NP NP 0.1 NP NP PP 0.2 NP N 0.7 PP P NP 1.0 VP  NP PP 1.0. N people 0. /6/7 CS 6/CS: Natural Language Processing Instructor: Prof. Lu Wang College of Computer and Information Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang The grammar: Binary, no epsilons,.9..5

More information

Transition-based dependency parsing

Transition-based dependency parsing Transition-based dependency parsing Daniël de Kok Overview Dependency graphs and treebanks. Transition-based dependency parsing. Making parse choices using perceptrons. Today Recap Transition systems Parsing

More information

Dependency Parsing. Statistical NLP Fall (Non-)Projectivity. CoNLL Format. Lecture 9: Dependency Parsing

Dependency Parsing. Statistical NLP Fall (Non-)Projectivity. CoNLL Format. Lecture 9: Dependency Parsing Dependency Parsing Statistical NLP Fall 2016 Lecture 9: Dependency Parsing Slav Petrov Google prep dobj ROOT nsubj pobj det PRON VERB DET NOUN ADP NOUN They solved the problem with statistics CoNLL Format

More information

Alessandro Mazzei MASTER DI SCIENZE COGNITIVE GENOVA 2005

Alessandro Mazzei MASTER DI SCIENZE COGNITIVE GENOVA 2005 Alessandro Mazzei Dipartimento di Informatica Università di Torino MATER DI CIENZE COGNITIVE GENOVA 2005 04-11-05 Natural Language Grammars and Parsing Natural Language yntax Paolo ama Francesca yntactic

More information

Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically

Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically NP S AUX VP Ali will V NP help D N the man A node is any point in the tree diagram and it can

More information

Natural Language Processing CS Lecture 06. Razvan C. Bunescu School of Electrical Engineering and Computer Science

Natural Language Processing CS Lecture 06. Razvan C. Bunescu School of Electrical Engineering and Computer Science Natural Language Processing CS 6840 Lecture 06 Razvan C. Bunescu School of Electrical Engineering and Computer Science bunescu@ohio.edu Statistical Parsing Define a probabilistic model of syntax P(T S):

More information

Penn Treebank Parsing. Advanced Topics in Language Processing Stephen Clark

Penn Treebank Parsing. Advanced Topics in Language Processing Stephen Clark Penn Treebank Parsing Advanced Topics in Language Processing Stephen Clark 1 The Penn Treebank 40,000 sentences of WSJ newspaper text annotated with phrasestructure trees The trees contain some predicate-argument

More information

Parsing with Context-Free Grammars

Parsing with Context-Free Grammars Parsing with Context-Free Grammars Berlin Chen 2005 References: 1. Natural Language Understanding, chapter 3 (3.1~3.4, 3.6) 2. Speech and Language Processing, chapters 9, 10 NLP-Berlin Chen 1 Grammars

More information

Maschinelle Sprachverarbeitung

Maschinelle Sprachverarbeitung Maschinelle Sprachverarbeitung Parsing with Probabilistic Context-Free Grammar Ulf Leser Content of this Lecture Phrase-Structure Parse Trees Probabilistic Context-Free Grammars Parsing with PCFG Other

More information

Maschinelle Sprachverarbeitung

Maschinelle Sprachverarbeitung Maschinelle Sprachverarbeitung Parsing with Probabilistic Context-Free Grammar Ulf Leser Content of this Lecture Phrase-Structure Parse Trees Probabilistic Context-Free Grammars Parsing with PCFG Other

More information

Transition-Based Parsing

Transition-Based Parsing Transition-Based Parsing Based on atutorial at COLING-ACL, Sydney 2006 with Joakim Nivre Sandra Kübler, Markus Dickinson Indiana University E-mail: skuebler,md7@indiana.edu Transition-Based Parsing 1(11)

More information

Parsing with Context-Free Grammars

Parsing with Context-Free Grammars Parsing with Context-Free Grammars CS 585, Fall 2017 Introduction to Natural Language Processing http://people.cs.umass.edu/~brenocon/inlp2017 Brendan O Connor College of Information and Computer Sciences

More information

Dependency Parsing. COSI 114 Computational Linguistics Marie Meteer. March 21, 2015 Brandeis University

Dependency Parsing. COSI 114 Computational Linguistics Marie Meteer. March 21, 2015 Brandeis University + Dependency Parsing COSI 114 Computational Linguistics Marie Meteer March 21, 2015 Brandeis University Dependency Grammar and Dependency Structure Dependency syntax postulates that syntac1c structure

More information

CS626: NLP, Speech and the Web. Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 14: Parsing Algorithms 30 th August, 2012

CS626: NLP, Speech and the Web. Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 14: Parsing Algorithms 30 th August, 2012 CS626: NLP, Speech and the Web Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 14: Parsing Algorithms 30 th August, 2012 Parsing Problem Semantics Part of Speech Tagging NLP Trinity Morph Analysis

More information

Personal Project: Shift-Reduce Dependency Parsing

Personal Project: Shift-Reduce Dependency Parsing Personal Project: Shift-Reduce Dependency Parsing 1 Problem Statement The goal of this project is to implement a shift-reduce dependency parser. This entails two subgoals: Inference: We must have a shift-reduce

More information

10/17/04. Today s Main Points

10/17/04. Today s Main Points Part-of-speech Tagging & Hidden Markov Model Intro Lecture #10 Introduction to Natural Language Processing CMPSCI 585, Fall 2004 University of Massachusetts Amherst Andrew McCallum Today s Main Points

More information

X-bar theory. X-bar :

X-bar theory. X-bar : is one of the greatest contributions of generative school in the filed of knowledge system. Besides linguistics, computer science is greatly indebted to Chomsky to have propounded the theory of x-bar.

More information

The Formal Architecture of. Lexical-Functional Grammar. Ronald M. Kaplan and Mary Dalrymple

The Formal Architecture of. Lexical-Functional Grammar. Ronald M. Kaplan and Mary Dalrymple The Formal Architecture of Lexical-Functional Grammar Ronald M. Kaplan and Mary Dalrymple Xerox Palo Alto Research Center 1. Kaplan and Dalrymple, ESSLLI 1995, Barcelona Architectural Issues Representation:

More information

LECTURER: BURCU CAN Spring

LECTURER: BURCU CAN Spring LECTURER: BURCU CAN 2017-2018 Spring Regular Language Hidden Markov Model (HMM) Context Free Language Context Sensitive Language Probabilistic Context Free Grammar (PCFG) Unrestricted Language PCFGs can

More information

Multiword Expression Identification with Tree Substitution Grammars

Multiword Expression Identification with Tree Substitution Grammars Multiword Expression Identification with Tree Substitution Grammars Spence Green, Marie-Catherine de Marneffe, John Bauer, and Christopher D. Manning Stanford University EMNLP 2011 Main Idea Use syntactic

More information

arxiv: v2 [cs.cl] 20 Apr 2017

arxiv: v2 [cs.cl] 20 Apr 2017 Syntax Aware LSM Model for Chinese Semantic Role Labeling Feng Qian 2, Lei Sha 1, Baobao Chang 1, Lu-chen Liu 2, Ming Zhang 2 1 Key Laboratory of Computational Linguistics, Ministry of Education, Peking

More information

Parsing with CFGs L445 / L545 / B659. Dept. of Linguistics, Indiana University Spring Parsing with CFGs. Direction of processing

Parsing with CFGs L445 / L545 / B659. Dept. of Linguistics, Indiana University Spring Parsing with CFGs. Direction of processing L445 / L545 / B659 Dept. of Linguistics, Indiana University Spring 2016 1 / 46 : Overview Input: a string Output: a (single) parse tree A useful step in the process of obtaining meaning We can view the

More information

Parsing with CFGs. Direction of processing. Top-down. Bottom-up. Left-corner parsing. Chart parsing CYK. Earley 1 / 46.

Parsing with CFGs. Direction of processing. Top-down. Bottom-up. Left-corner parsing. Chart parsing CYK. Earley 1 / 46. : Overview L545 Dept. of Linguistics, Indiana University Spring 2013 Input: a string Output: a (single) parse tree A useful step in the process of obtaining meaning We can view the problem as searching

More information

Spatial Role Labeling CS365 Course Project

Spatial Role Labeling CS365 Course Project Spatial Role Labeling CS365 Course Project Amit Kumar, akkumar@iitk.ac.in Chandra Sekhar, gchandra@iitk.ac.in Supervisor : Dr.Amitabha Mukerjee ABSTRACT In natural language processing one of the important

More information

Model-Theory of Property Grammars with Features

Model-Theory of Property Grammars with Features Model-Theory of Property Grammars with Features Denys Duchier Thi-Bich-Hanh Dao firstname.lastname@univ-orleans.fr Yannick Parmentier Abstract In this paper, we present a model-theoretic description of

More information

Bringing machine learning & compositional semantics together: central concepts

Bringing machine learning & compositional semantics together: central concepts Bringing machine learning & compositional semantics together: central concepts https://githubcom/cgpotts/annualreview-complearning Chris Potts Stanford Linguistics CS 244U: Natural language understanding

More information

NLP Homework: Dependency Parsing with Feed-Forward Neural Network

NLP Homework: Dependency Parsing with Feed-Forward Neural Network NLP Homework: Dependency Parsing with Feed-Forward Neural Network Submission Deadline: Monday Dec. 11th, 5 pm 1 Background on Dependency Parsing Dependency trees are one of the main representations used

More information

CS460/626 : Natural Language

CS460/626 : Natural Language CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 23, 24 Parsing Algorithms; Parsing in case of Ambiguity; Probabilistic Parsing) Pushpak Bhattacharyya CSE Dept., IIT Bombay 8 th,

More information

13A. Computational Linguistics. 13A. Log-Likelihood Dependency Parsing. CSC 2501 / 485 Fall 2017

13A. Computational Linguistics. 13A. Log-Likelihood Dependency Parsing. CSC 2501 / 485 Fall 2017 Computational Linguistics CSC 2501 / 485 Fall 2017 13A 13A. Log-Likelihood Dependency Parsing Gerald Penn Department of Computer Science, University of Toronto Based on slides by Yuji Matsumoto, Dragomir

More information

Parsing. Based on presentations from Chris Manning s course on Statistical Parsing (Stanford)

Parsing. Based on presentations from Chris Manning s course on Statistical Parsing (Stanford) Parsing Based on presentations from Chris Manning s course on Statistical Parsing (Stanford) S N VP V NP D N John hit the ball Levels of analysis Level Morphology/Lexical POS (morpho-synactic), WSD Elements

More information

Natural Language Processing

Natural Language Processing SFU NatLangLab Natural Language Processing Anoop Sarkar anoopsarkar.github.io/nlp-class Simon Fraser University September 27, 2018 0 Natural Language Processing Anoop Sarkar anoopsarkar.github.io/nlp-class

More information

Extensions to the Logic of All x are y: Verbs, Relative Clauses, and Only

Extensions to the Logic of All x are y: Verbs, Relative Clauses, and Only 1/53 Extensions to the Logic of All x are y: Verbs, Relative Clauses, and Only Larry Moss Indiana University Nordic Logic School August 7-11, 2017 2/53 An example that we ll see a few times Consider the

More information

Probabilistic Context-free Grammars

Probabilistic Context-free Grammars Probabilistic Context-free Grammars Computational Linguistics Alexander Koller 24 November 2017 The CKY Recognizer S NP VP NP Det N VP V NP V ate NP John Det a N sandwich i = 1 2 3 4 k = 2 3 4 5 S NP John

More information

A* Search. 1 Dijkstra Shortest Path

A* Search. 1 Dijkstra Shortest Path A* Search Consider the eight puzzle. There are eight tiles numbered 1 through 8 on a 3 by three grid with nine locations so that one location is left empty. We can move by sliding a tile adjacent to the

More information

The SUBTLE NL Parsing Pipeline: A Complete Parser for English Mitch Marcus University of Pennsylvania

The SUBTLE NL Parsing Pipeline: A Complete Parser for English Mitch Marcus University of Pennsylvania The SUBTLE NL Parsing Pipeline: A Complete Parser for English Mitch Marcus University of Pennsylvania 1 PICTURE OF ANALYSIS PIPELINE Tokenize Maximum Entropy POS tagger MXPOST Ratnaparkhi Core Parser Collins

More information

Models of Adjunction in Minimalist Grammars

Models of Adjunction in Minimalist Grammars Models of Adjunction in Minimalist Grammars Thomas Graf mail@thomasgraf.net http://thomasgraf.net Stony Brook University FG 2014 August 17, 2014 The Theory-Neutral CliffsNotes Insights Several properties

More information

A DOP Model for LFG. Rens Bod and Ronald Kaplan. Kathrin Spreyer Data-Oriented Parsing, 14 June 2005

A DOP Model for LFG. Rens Bod and Ronald Kaplan. Kathrin Spreyer Data-Oriented Parsing, 14 June 2005 A DOP Model for LFG Rens Bod and Ronald Kaplan Kathrin Spreyer Data-Oriented Parsing, 14 June 2005 Lexical-Functional Grammar (LFG) Levels of linguistic knowledge represented formally differently (non-monostratal):

More information

Artificial Intelligence

Artificial Intelligence CS344: Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 20-21 Natural Language Parsing Parsing of Sentences Are sentences flat linear structures? Why tree? Is

More information

A Syntax-based Statistical Machine Translation Model. Alexander Friedl, Georg Teichtmeister

A Syntax-based Statistical Machine Translation Model. Alexander Friedl, Georg Teichtmeister A Syntax-based Statistical Machine Translation Model Alexander Friedl, Georg Teichtmeister 4.12.2006 Introduction The model Experiment Conclusion Statistical Translation Model (STM): - mathematical model

More information

Word Embeddings in Feedforward Networks; Tagging and Dependency Parsing using Feedforward Networks. Michael Collins, Columbia University

Word Embeddings in Feedforward Networks; Tagging and Dependency Parsing using Feedforward Networks. Michael Collins, Columbia University Word Embeddings in Feedforward Networks; Tagging and Dependency Parsing using Feedforward Networks Michael Collins, Columbia University Overview Introduction Multi-layer feedforward networks Representing

More information

Natural Language Processing 1. lecture 7: constituent parsing. Ivan Titov. Institute for Logic, Language and Computation

Natural Language Processing 1. lecture 7: constituent parsing. Ivan Titov. Institute for Logic, Language and Computation atural Language Processing 1 lecture 7: constituent parsing Ivan Titov Institute for Logic, Language and Computation Outline Syntax: intro, CFGs, PCFGs PCFGs: Estimation CFGs: Parsing PCFGs: Parsing Parsing

More information

Context-free grammars for natural languages

Context-free grammars for natural languages Context-free grammars Basic definitions Context-free grammars for natural languages Context-free grammars can be used for a variety of syntactic constructions, including some non-trivial phenomena such

More information

The Real Story on Synchronous Rewriting Systems. Daniel Gildea Computer Science Department University of Rochester

The Real Story on Synchronous Rewriting Systems. Daniel Gildea Computer Science Department University of Rochester The Real Story on Synchronous Rewriting Systems Daniel Gildea omputer Science Department University of Rochester Overview Factorization Tree Decomposition Factorization of LFRS YK Parsing w 2 : [, x 1,

More information

Ling 240 Lecture #15. Syntax 4

Ling 240 Lecture #15. Syntax 4 Ling 240 Lecture #15 Syntax 4 agenda for today Give me homework 3! Language presentation! return Quiz 2 brief review of friday More on transformation Homework 4 A set of Phrase Structure Rules S -> (aux)

More information

Parsing Beyond Context-Free Grammars: Tree Adjoining Grammars

Parsing Beyond Context-Free Grammars: Tree Adjoining Grammars Parsing Beyond Context-Free Grammars: Tree Adjoining Grammars Laura Kallmeyer & Tatiana Bladier Heinrich-Heine-Universität Düsseldorf Sommersemester 2018 Kallmeyer, Bladier SS 2018 Parsing Beyond CFG:

More information

TALP at GeoQuery 2007: Linguistic and Geographical Analysis for Query Parsing

TALP at GeoQuery 2007: Linguistic and Geographical Analysis for Query Parsing TALP at GeoQuery 2007: Linguistic and Geographical Analysis for Query Parsing Daniel Ferrés and Horacio Rodríguez TALP Research Center Software Department Universitat Politècnica de Catalunya {dferres,horacio}@lsi.upc.edu

More information

Posterior vs. Parameter Sparsity in Latent Variable Models Supplementary Material

Posterior vs. Parameter Sparsity in Latent Variable Models Supplementary Material Posterior vs. Parameter Sparsity in Latent Variable Models Supplementary Material João V. Graça L 2 F INESC-ID Lisboa, Portugal Kuzman Ganchev Ben Taskar University of Pennsylvania Philadelphia, PA, USA

More information

(7) a. [ PP to John], Mary gave the book t [PP]. b. [ VP fix the car], I wonder whether she will t [VP].

(7) a. [ PP to John], Mary gave the book t [PP]. b. [ VP fix the car], I wonder whether she will t [VP]. CAS LX 522 Syntax I Fall 2000 September 18, 2000 Paul Hagstrom Week 2: Movement Movement Last time, we talked about subcategorization. (1) a. I can solve this problem. b. This problem, I can solve. (2)

More information

Modular Grammar Design with Typed Parametric Principles

Modular Grammar Design with Typed Parametric Principles 1 Modular Grammar Design with Typed Parametric Principles 1.1 Introduction Extensible Dependency Grammar (xdg) (Debusmann et al., 2004) is a general framework for dependency grammar, with multiple levels

More information

National Centre for Language Technology School of Computing Dublin City University

National Centre for Language Technology School of Computing Dublin City University with with National Centre for Language Technology School of Computing Dublin City University Parallel treebanks A parallel treebank comprises: sentence pairs parsed word-aligned tree-aligned (Volk & Samuelsson,

More information

Spectral Unsupervised Parsing with Additive Tree Metrics

Spectral Unsupervised Parsing with Additive Tree Metrics Spectral Unsupervised Parsing with Additive Tree Metrics Ankur Parikh, Shay Cohen, Eric P. Xing Carnegie Mellon, University of Edinburgh Ankur Parikh 2014 1 Overview Model: We present a novel approach

More information

Department of Computer Science and Engineering Indian Institute of Technology, Kanpur. Spatial Role Labeling

Department of Computer Science and Engineering Indian Institute of Technology, Kanpur. Spatial Role Labeling Department of Computer Science and Engineering Indian Institute of Technology, Kanpur CS 365 Artificial Intelligence Project Report Spatial Role Labeling Submitted by Satvik Gupta (12633) and Garvit Pahal

More information

1. Background. Task: Determine whether a given string of words is a grammatical (well-formed) sentence of language L i or not.

1. Background. Task: Determine whether a given string of words is a grammatical (well-formed) sentence of language L i or not. Constraints in Syntax [1] Phrase Structure and Derivations Düsseldorf LSA/DGfS Summerschool 2002 Gereon Müller (IDS Mannheim) gereon.mueller@ids-mannheim.de 1. Background Task: Determine whether a given

More information

POS-Tagging. Fabian M. Suchanek

POS-Tagging. Fabian M. Suchanek POS-Tagging Fabian M. Suchanek 100 Def: POS A Part-of-Speech (also: POS, POS-tag, word class, lexical class, lexical category) is a set of words with the same grammatical role. Alizée wrote a really great

More information

Part-of-Speech Tagging + Neural Networks CS 287

Part-of-Speech Tagging + Neural Networks CS 287 Part-of-Speech Tagging + Neural Networks CS 287 Quiz Last class we focused on hinge loss. L hinge = max{0, 1 (ŷ c ŷ c )} Consider now the squared hinge loss, (also called l 2 SVM) L hinge 2 = max{0, 1

More information

CS 188 Introduction to AI Fall 2005 Stuart Russell Final

CS 188 Introduction to AI Fall 2005 Stuart Russell Final NAME: SID#: Section: 1 CS 188 Introduction to AI all 2005 Stuart Russell inal You have 2 hours and 50 minutes. he exam is open-book, open-notes. 100 points total. Panic not. Mark your answers ON HE EXAM

More information

Parasitic Scope (Barker 2007) Semantics Seminar 11/10/08

Parasitic Scope (Barker 2007) Semantics Seminar 11/10/08 Parasitic Scope (Barker 2007) Semantics Seminar 11/10/08 1. Overview Attempts to provide a compositional, fully semantic account of same. Elements other than NPs in particular, adjectives can be scope-taking

More information

Probabilistic Context Free Grammars. Many slides from Michael Collins

Probabilistic Context Free Grammars. Many slides from Michael Collins Probabilistic Context Free Grammars Many slides from Michael Collins Overview I Probabilistic Context-Free Grammars (PCFGs) I The CKY Algorithm for parsing with PCFGs A Probabilistic Context-Free Grammar

More information

HPSG: Binding Theory

HPSG: Binding Theory HPSG: Binding Theory Doug Arnold doug@essexacuk Introduction Binding Theory is to do with the syntactic restrictions on the distribution of referentially dependent items and their antecedents: reflexives/reciprocals

More information

A Context-Free Grammar

A Context-Free Grammar Statistical Parsing A Context-Free Grammar S VP VP Vi VP Vt VP VP PP DT NN PP PP P Vi sleeps Vt saw NN man NN dog NN telescope DT the IN with IN in Ambiguity A sentence of reasonable length can easily

More information

Lecture 5: UDOP, Dependency Grammars

Lecture 5: UDOP, Dependency Grammars Lecture 5: UDOP, Dependency Grammars Jelle Zuidema ILLC, Universiteit van Amsterdam Unsupervised Language Learning, 2014 Generative Model objective PCFG PTSG CCM DMV heuristic Wolff (1984) UDOP ML IO K&M

More information

CS 6120/CS4120: Natural Language Processing

CS 6120/CS4120: Natural Language Processing CS 6120/CS4120: Natural Language Processing Instructor: Prof. Lu Wang College of Computer and Information Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Assignment/report submission

More information

Introduction to Semantics. The Formalization of Meaning 1

Introduction to Semantics. The Formalization of Meaning 1 The Formalization of Meaning 1 1. Obtaining a System That Derives Truth Conditions (1) The Goal of Our Enterprise To develop a system that, for every sentence S of English, derives the truth-conditions

More information

Two-phase Implementation of Morphological Analysis

Two-phase Implementation of Morphological Analysis Two-phase Implementation of Morphological Analysis Arvi Hurskainen Institute for Asian and African Studies, Box 59 FIN-00014 University of Helsinki, Finland arvi.hurskainen@helsinki.fi Abstract SALAMA

More information

Part of Speech Tagging: Viterbi, Forward, Backward, Forward- Backward, Baum-Welch. COMP-599 Oct 1, 2015

Part of Speech Tagging: Viterbi, Forward, Backward, Forward- Backward, Baum-Welch. COMP-599 Oct 1, 2015 Part of Speech Tagging: Viterbi, Forward, Backward, Forward- Backward, Baum-Welch COMP-599 Oct 1, 2015 Announcements Research skills workshop today 3pm-4:30pm Schulich Library room 313 Start thinking about

More information

Design and Evaluation of the Linguistic Basis of an Automatic F-Structure Annotation Algorithm for the Penn-II Treebank

Design and Evaluation of the Linguistic Basis of an Automatic F-Structure Annotation Algorithm for the Penn-II Treebank Design and Evaluation of the Linguistic Basis of an Automatic F-Structure Annotation Algorithm for the Penn-II Treebank Mairead McCarthy M Sc m Computer Applications, School of Computing, Dublin City University

More information

Statistical Methods for NLP

Statistical Methods for NLP Statistical Methods for NLP Stochastic Grammars Joakim Nivre Uppsala University Department of Linguistics and Philology joakim.nivre@lingfil.uu.se Statistical Methods for NLP 1(22) Structured Classification

More information

Handout 8: Computation & Hierarchical parsing II. Compute initial state set S 0 Compute initial state set S 0

Handout 8: Computation & Hierarchical parsing II. Compute initial state set S 0 Compute initial state set S 0 Massachusetts Institute of Technology 6.863J/9.611J, Natural Language Processing, Spring, 2001 Department of Electrical Engineering and Computer Science Department of Brain and Cognitive Sciences Handout

More information

Structured Prediction Models via the Matrix-Tree Theorem

Structured Prediction Models via the Matrix-Tree Theorem Structured Prediction Models via the Matrix-Tree Theorem Terry Koo Amir Globerson Xavier Carreras Michael Collins maestro@csail.mit.edu gamir@csail.mit.edu carreras@csail.mit.edu mcollins@csail.mit.edu

More information

INF4820: Algorithms for Artificial Intelligence and Natural Language Processing. Hidden Markov Models

INF4820: Algorithms for Artificial Intelligence and Natural Language Processing. Hidden Markov Models INF4820: Algorithms for Artificial Intelligence and Natural Language Processing Hidden Markov Models Murhaf Fares & Stephan Oepen Language Technology Group (LTG) October 27, 2016 Recap: Probabilistic Language

More information

Other types of Movement

Other types of Movement Other types of Movement So far we seen Wh-movement, which moves certain types of (XP) constituents to the specifier of a CP. Wh-movement is also called A-bar movement. We will look at two more types of

More information

Splitting the structured paths in stratified graphs. Application in Natural Language Generation

Splitting the structured paths in stratified graphs. Application in Natural Language Generation DOI: 10.2478/auom-2014-0031 An. Şt. Univ. Ovidius Constanţa Vol. 22(2),2014, 57 67 Splitting the structured paths in stratified graphs. Application in Natural Language Generation Dana Dănciulescu and Mihaela

More information

Ling 5801: Lecture Notes 7 From Programs to Context-Free Grammars

Ling 5801: Lecture Notes 7 From Programs to Context-Free Grammars Ling 5801: Lecture otes 7 From Programs to Context-Free rammars 1. The rules we used to define programs make up a context-free grammar A Context-Free rammar is a tuple C,X,S,R, where: C is a finite set

More information

Syntax-Based Decoding

Syntax-Based Decoding Syntax-Based Decoding Philipp Koehn 9 November 2017 1 syntax-based models Synchronous Context Free Grammar Rules 2 Nonterminal rules NP DET 1 2 JJ 3 DET 1 JJ 3 2 Terminal rules N maison house NP la maison

More information

Constituency. Doug Arnold

Constituency. Doug Arnold Constituency Doug Arnold doug@essex.ac.uk Spose we have a string... xyz..., how can we establish whether xyz is a constituent (i.e. syntactic unit); i.e. whether the representation of... xyz... should

More information

Proseminar on Semantic Theory Fall 2010 Ling 720. Remko Scha (1981/1984): Distributive, Collective and Cumulative Quantification

Proseminar on Semantic Theory Fall 2010 Ling 720. Remko Scha (1981/1984): Distributive, Collective and Cumulative Quantification 1. Introduction Remko Scha (1981/1984): Distributive, Collective and Cumulative Quantification (1) The Importance of Scha (1981/1984) The first modern work on plurals (Landman 2000) There are many ideas

More information

CS 545 Lecture XVI: Parsing

CS 545 Lecture XVI: Parsing CS 545 Lecture XVI: Parsing brownies_choco81@yahoo.com brownies_choco81@yahoo.com Benjamin Snyder Parsing Given a grammar G and a sentence x = (x1, x2,..., xn), find the best parse tree. We re not going

More information

Annotation tasks and solutions in CLARIN-PL

Annotation tasks and solutions in CLARIN-PL Annotation tasks and solutions in CLARIN-PL Marcin Oleksy, Ewa Rudnicka Wrocław University of Technology marcin.oleksy@pwr.edu.pl ewa.rudnicka@pwr.edu.pl CLARIN ERIC Common Language Resources and Technology

More information

CMPT-825 Natural Language Processing. Why are parsing algorithms important?

CMPT-825 Natural Language Processing. Why are parsing algorithms important? CMPT-825 Natural Language Processing Anoop Sarkar http://www.cs.sfu.ca/ anoop October 26, 2010 1/34 Why are parsing algorithms important? A linguistic theory is implemented in a formal system to generate

More information

A Supertag-Context Model for Weakly-Supervised CCG Parser Learning

A Supertag-Context Model for Weakly-Supervised CCG Parser Learning A Supertag-Context Model for Weakly-Supervised CCG Parser Learning Dan Garrette Chris Dyer Jason Baldridge Noah A. Smith U. Washington CMU UT-Austin CMU Contributions 1. A new generative model for learning

More information

Unterspezifikation in der Semantik Scope Semantics in Lexicalized Tree Adjoining Grammars

Unterspezifikation in der Semantik Scope Semantics in Lexicalized Tree Adjoining Grammars in der emantik cope emantics in Lexicalized Tree Adjoining Grammars Laura Heinrich-Heine-Universität Düsseldorf Wintersemester 2011/2012 LTAG: The Formalism (1) Tree Adjoining Grammars (TAG): Tree-rewriting

More information

Treebank Grammar Techniques for Non-Projective Dependency Parsing

Treebank Grammar Techniques for Non-Projective Dependency Parsing Treebank Grammar Techniques for Non-Projective Dependency Parsing Marco Kuhlmann Uppsala University Uppsala, Sweden marco.kuhlmann@lingfil.uu.se Giorgio Satta University of Padua Padova, Italy satta@dei.unipd.it

More information

Categorial Grammar. Larry Moss NASSLLI. Indiana University

Categorial Grammar. Larry Moss NASSLLI. Indiana University 1/37 Categorial Grammar Larry Moss Indiana University NASSLLI 2/37 Categorial Grammar (CG) CG is the tradition in grammar that is closest to the work that we ll do in this course. Reason: In CG, syntax

More information

Effectiveness of complex index terms in information retrieval

Effectiveness of complex index terms in information retrieval Effectiveness of complex index terms in information retrieval Tokunaga Takenobu, Ogibayasi Hironori and Tanaka Hozumi Department of Computer Science Tokyo Institute of Technology Abstract This paper explores

More information

Grammar and Feature Unification

Grammar and Feature Unification Grammar and Feature Unification Problems with CF Phrase Structure Grammars Difficult to capture dependencies between constituents the boy runs the boys run * the boy run * the boys runs Problems with CF

More information

Graph-based Dependency Parsing. Ryan McDonald Google Research

Graph-based Dependency Parsing. Ryan McDonald Google Research Graph-based Dependency Parsing Ryan McDonald Google Research ryanmcd@google.com Reader s Digest Graph-based Dependency Parsing Ryan McDonald Google Research ryanmcd@google.com root ROOT Dependency Parsing

More information

Marrying Dynamic Programming with Recurrent Neural Networks

Marrying Dynamic Programming with Recurrent Neural Networks Marrying Dynamic Programming with Recurrent Neural Networks I eat sushi with tuna from Japan Liang Huang Oregon State University Structured Prediction Workshop, EMNLP 2017, Copenhagen, Denmark Marrying

More information

Wide-coverage Translation in GF

Wide-coverage Translation in GF Wide-coverage Translation in GF Krasimir Angelov University of Gothenburg, Digital Grammars AB August 18, 2017 1 Introduction 2 Free Language Resource Grammar Resource Lexicon Demo 3 Uncontrolled Language

More information

Natural Language Processing : Probabilistic Context Free Grammars. Updated 5/09

Natural Language Processing : Probabilistic Context Free Grammars. Updated 5/09 Natural Language Processing : Probabilistic Context Free Grammars Updated 5/09 Motivation N-gram models and HMM Tagging only allowed us to process sentences linearly. However, even simple sentences require

More information

Review. Earley Algorithm Chapter Left Recursion. Left-Recursion. Rule Ordering. Rule Ordering

Review. Earley Algorithm Chapter Left Recursion. Left-Recursion. Rule Ordering. Rule Ordering Review Earley Algorithm Chapter 13.4 Lecture #9 October 2009 Top-Down vs. Bottom-Up Parsers Both generate too many useless trees Combine the two to avoid over-generation: Top-Down Parsing with Bottom-Up

More information

Vine Pruning for Efficient Multi-Pass Dependency Parsing. Alexander M. Rush and Slav Petrov

Vine Pruning for Efficient Multi-Pass Dependency Parsing. Alexander M. Rush and Slav Petrov Vine Pruning for Efficient Multi-Pass Dependency Parsing Alexander M. Rush and Slav Petrov Dependency Parsing Styles of Dependency Parsing greedy O(n) transition-based parsers (Nivre 2004) graph-based

More information

Compositionality and Syntactic Structure Marcus Kracht Department of Linguistics UCLA 3125 Campbell Hall 405 Hilgard Avenue Los Angeles, CA 90095

Compositionality and Syntactic Structure Marcus Kracht Department of Linguistics UCLA 3125 Campbell Hall 405 Hilgard Avenue Los Angeles, CA 90095 Compositionality and Syntactic Structure Marcus Kracht Department of Linguistics UCLA 3125 Campbell Hall 405 Hilgard Avenue Los Angeles, CA 90095 1543 kracht@humnet.ucla.edu 1. The Questions ➀ Why does

More information

Seminar in Semantics: Gradation & Modality Winter 2014

Seminar in Semantics: Gradation & Modality Winter 2014 1 Subject matter Seminar in Semantics: Gradation & Modality Winter 2014 Dan Lassiter 1/8/14 Handout: Basic Modal Logic and Kratzer (1977) [M]odality is the linguistic phenomenon whereby grammar allows

More information

Modular Grammar Design with Typed Parametric Principles

Modular Grammar Design with Typed Parametric Principles 11 Modular Grammar Design with Typed Parametric Principles Ralph Debusmann, Denys Duchier and Andreas Rossberg Abstract This paper introduces a type system for Extensible Dependency Grammar (xdg) (Debusmann

More information

C SC 620 Advanced Topics in Natural Language Processing. Lecture 21 4/13

C SC 620 Advanced Topics in Natural Language Processing. Lecture 21 4/13 C SC 620 Advanced Topics in Natural Language Processing Lecture 21 4/13 Reading List Readings in Machine Translation, Eds. Nirenburg, S. et al. MIT Press 2003. 19. Montague Grammar and Machine Translation.

More information

Simpler Syntax. Ling : Sign-Based Construction Grammar Instructor: Ivan A. Sag URL:

Simpler Syntax. Ling : Sign-Based Construction Grammar Instructor: Ivan A. Sag URL: Simpler Syntax Ling 7800-065: Sign-Based Construction Grammar Instructor: Ivan A. Sag (sag@stanford.edu) URL: http://lingo.stanford.edu/sag/li11-sbcg 1/ 58 Constructs ] [mtr sign construct : dtrs nelist(sign)

More information

Adaptive Multi-Compositionality for Recursive Neural Models with Applications to Sentiment Analysis. July 31, 2014

Adaptive Multi-Compositionality for Recursive Neural Models with Applications to Sentiment Analysis. July 31, 2014 Adaptive Multi-Compositionality for Recursive Neural Models with Applications to Sentiment Analysis July 31, 2014 Semantic Composition Principle of Compositionality The meaning of a complex expression

More information

Unification. Two Routes to Deep Structure. Unification. Unification Grammar. Martin Kay. Stanford University University of the Saarland

Unification. Two Routes to Deep Structure. Unification. Unification Grammar. Martin Kay. Stanford University University of the Saarland Two Routes to Deep Structure Derivational! Transformational! Procedural Deep Unidirectional Transformations Surface Stanford University University of the Saarland Constraint-based! Declarative Deep Surface

More information