Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically
|
|
- Norah Brooks
- 6 years ago
- Views:
Transcription
1 Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically NP S AUX VP Ali will V NP help D N the man A node is any point in the tree diagram and it can be: Branching node like S and lower NP. Non-branching node like AUX and V. Terminal nodes: lexical items at end of tree like help Nodes are related to each other by two relations:
2 1. Dominance: A node X dominates node Y if: X is higher than Y and X is connected to Y by a branch. For example, NP dominates Ali, VP dominates D and NP. Immediate dominance: A node immediately dominates another if there s s no intervening node. e.g. S immediately dominates NP, AUX, VP but not help. 2. Precedence: A node X precedes Y if it is on the left and both aren t dominating each other. e.g. Ali precedes will,, but the doesn t precede man.
3 Phrase Structure Rules These rules derive different types of phrases and unlimited number of sentences: NP (D) (AdjP( AdjP) ) N This rule cannot predict the structure of the phrase and we need to memorize infinite number of rules. The Structure of Phrases 1. VP: It consists of - lexical category: head V - Phrasal category or maximal projection: : VP as a whole - Intermediate category: V V (part of VP)
4 V V is a level that contains the object and verb s s modifiers: The head verb (and its object if there is one) is in the lowest V level The verb s s modifiers are placed in higher V V levels, called adjunct Subject combines with highest V V level, i.e. specifier of VP For example: they eat lunch in school tonight VP spec V V they V V NP V PP tonight V NP in school eat lunch
5 Evidence for V V V is a constituent structure that can be replaced by do so: - They eat lunch in school tonight and Ali does so. (eat lunch in school tonight) - They eat lunch in school tonight and Ali did so this morning. (eat lunch in school) - They eat lunch in school tonight and Ali does so at work this morning. (eat lunch) adjuncts are recursive, i.e. repeatedly added. 2. NP N is a level that contains the object and noun s s modifiers: The head noun (and its object if there is one) is in the lowest N N level The noun s s modifiers are placed in higher N N levels, called adjunct The determiner combines with the highest N, N, i.e. specifier of VP For example, the big book of poems with the blue cover
6 NP Det N 1 the N 2N AP N 3 N 3 PP big N PP with blue cover book of poems Evidence for N N N is a constituent structure that can be replaced by one: - I want this [big book of poems with the blue cover] not that one (N 1) - I want this big [book of poems with the blue cover] not that small one (N 2) - I want this big [book of poems] with the blue cover not that small one with the red cover (N 3) 3. AdjP & PP We apply the same structure to these phrases:
7 He is [quite jealous of Ali]. He stood [right across the bridge] The lowest P P and adj includes these heads (and their complements). The spec combines with P P and adj and is placed by modifiers (like quite, very, rather, so for the adj and straight, right for the prep). General structure (x-bar) We can have these general rules that exactly predict the structure re of different types of phrases: XP spec X X (specifier) X X YP (head and complement) X X YP (adjunct= modifier)
8 Sentence structure Aux is the head of the sentence because it carries tense and agreement: They are working hard. But what about non-auxiliary verbs: They worked hard Aux is the head even though it isn t t overt. Evidence for aux as head of sentence: 1. Cleft sentence: work hard, they did indeed. The tense is on aux and not part of VP.
9 2. Pseudo-cleft: what they did was work hard. Tense is part of a node, Infl(ection),, which can be filled with overt aux or left empty. s NP I VP +tense +agr I is finite because it has (+t, +agr+ agr). Infinitive clause: I ask [Ali to work hard] To is the head of the infin clause and it s -t, -agr, non-finite I
10 The structure of IP IP NP I I VP will V V -ed V NP to finish the work IP spec I I I I VP VP is always a complement of I.
11 IP is a functional category, not a lexical category, because it is used for grammatical function: t, agr Complementizer phrase (CP) CP is another functional category since its head, c, introduces a subordinate clause: C= that, for [-wh[ wh] if, whether [+wh wh] I believe [that Ali will work hard]. I want [for Ali to work hard]. I wonder [if/whether Ali worked hard].
12 The above CP clauses have the following structure: CP spec C C C IP Ali will worked hard - IP is always the complement of C. - C is filled by that, for and moved will,, forming yes/no question: CP spec C C C IP will NP I [+wh wh] ] Ali I VP work hard
13 - C can be filled by either will or that, not both - Spec is filled by whether, if and wh-questions: Structure of CP: CP Spec C C C C IP CP spec C C when C IP will NP I Ali I VP work hard
14 Structural relations Government: A head governs its phrase e.g. N heads its phrase, I the IP, and C the CP. Agreement between head and non-head is established under government: 1. NP agreement NP Det N this/these N PP car/cars of Ali
15 There s s agreement in number between head (N) and spec (det). But agreement is poor in English since it s s shown morphologically in number not in gender and person as in French. 2. sub/ verb agreement There spec-head agreement between Infl and spec (sub) in number, person: IP spec I I He/they I VP -s s / ø play
16 3. CP agreement There s s a spec-head agreement in CP in [wh[ wh] ] feature: I wonder if he played. CP spec C C if C IP [+wh wh] ] [+wh wh] ] played C-command A c-commands c commands B if and if i. A doesn t t dominate B and B doesn t t dominate A; and ii. The first branching node dominating A dominates B. For instance, in (2) the spec c-commands commands every node in IP, but VP only I.
17 Let us now refine the notion of government: Government: A governs B if and if: A is a governor (i.e. a head) and A c-commands c commands B. Strict c-command c command & M-commandM 1. Strict c-commandc command VP V V PP V NP P P NP ate the food in the garden V strictly c-commands c commands NP because they re dominated by 1 st branching node but not PP.
18 2. M-commandM The head c-commands c commands the adjunct in the garden since there is a node VP that dominate both of them: A c-commands c commands B if and only if: A doesn t t dominate B and every node X (maximal projection) dominating A dominates B. V c-commands c commands NP, V doesn t t c-commands c commands PP (dominance by1 st branching node) V m-commands m NP, PP, NP (inside PP) (dominance by maximal proj.) P doesn t t m-commands m V (PP doesn t t dominate V)
19 Government is refined as: A governs B if and if: i. A is a governor; and ii. A m-commands m B; and iii. No barrier between A and B. In (1), the verb ate governs the PP, but it doesn t t govern the NP since it is a barrier. However the verb m-commands m NP.
X-bar theory. X-bar :
is one of the greatest contributions of generative school in the filed of knowledge system. Besides linguistics, computer science is greatly indebted to Chomsky to have propounded the theory of x-bar.
More informationParsing with Context-Free Grammars
Parsing with Context-Free Grammars Berlin Chen 2005 References: 1. Natural Language Understanding, chapter 3 (3.1~3.4, 3.6) 2. Speech and Language Processing, chapters 9, 10 NLP-Berlin Chen 1 Grammars
More information(7) a. [ PP to John], Mary gave the book t [PP]. b. [ VP fix the car], I wonder whether she will t [VP].
CAS LX 522 Syntax I Fall 2000 September 18, 2000 Paul Hagstrom Week 2: Movement Movement Last time, we talked about subcategorization. (1) a. I can solve this problem. b. This problem, I can solve. (2)
More informationCAS LX 522 Syntax I Fall 2000 October 10, 2000 Week 5: Case Theory and θ Theory. θ-theory continued
CAS LX 522 Syntax I Fall 2000 October 0, 2000 Paul Hagstrom Week 5: Case Theory and θ Theory θ-theory continued From last time: verbs have θ-roles (e.g., Agent, Theme, ) to assign, specified in the lexicon
More informationAnother look at PSRs: Intermediate Structure. Starting X-bar theory
Another look at PSRs: Intermediate Structure Starting X-bar theory Andrew Carnie, 2006 Substitution Andrew Carnie, 2006 Substitution If a group of words can be replaced by a single word, they are a constituent.
More informationComputationele grammatica
Computationele grammatica Docent: Paola Monachesi Contents First Last Prev Next Contents 1 Unbounded dependency constructions (UDCs)............... 3 2 Some data...............................................
More information1. Background. Task: Determine whether a given string of words is a grammatical (well-formed) sentence of language L i or not.
Constraints in Syntax [1] Phrase Structure and Derivations Düsseldorf LSA/DGfS Summerschool 2002 Gereon Müller (IDS Mannheim) gereon.mueller@ids-mannheim.de 1. Background Task: Determine whether a given
More informationOther types of Movement
Other types of Movement So far we seen Wh-movement, which moves certain types of (XP) constituents to the specifier of a CP. Wh-movement is also called A-bar movement. We will look at two more types of
More informationLing 240 Lecture #15. Syntax 4
Ling 240 Lecture #15 Syntax 4 agenda for today Give me homework 3! Language presentation! return Quiz 2 brief review of friday More on transformation Homework 4 A set of Phrase Structure Rules S -> (aux)
More informationArtificial Intelligence
CS344: Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 20-21 Natural Language Parsing Parsing of Sentences Are sentences flat linear structures? Why tree? Is
More informationCS 712: Topics in NLP Linguistic Phrases and Statistical Phrases
CS 712: Topics in NLP Linguistic Phrases and Statistical Phrases Pushpak Bhattacharyya, CSE Department, IIT Bombay 18 March, 2013 (main text: Syntax by Adrew Carnie, Blackwell Publication, 2002) Domination
More informationCAS LX 523 Syntax II Spring 2001 March 13, (1) A qp. Kayne, Richard (1995). The antisymmetry of syntax. Cambridge, MA: MIT Press.
CAS LX 52 Syntax II Spring 200 March, 200 Paul Hagstrom Week 7: Antisymmetry Kayne, Richard (995). The antisymmetry of syntax. Cambridge, MA: MIT Press. Koopman, Hilda (2000). The spec-head configuration.
More information2 A not-quite-argument for X-bar structure in noun phrases
CAS LX 321 / GRS LX 621 Syntax: Introduction to Sentential Structure ovember 16, 2017 1 and pronouns (1) he linguists yodel. (2) We linguists yodel. (3) hey looked at us linguists. (4) hey looked at linguists.
More informationAndrew Carnie, Structural Relations. The mathematical properties of phrase structure trees
Structural Relations The mathematical properties of phrase structure trees Important! Important! Even if you have trouble with the formal definitions, try to understand the INTUITIVE idea behind them.
More informationTHE DRAVIDIAN EXPERIENCER CONSTRUCTION AND THE ENGLISH SEEM CONSTRUCTION. K. A. Jayaseelan CIEFL, Hyderabad
THE DRAVIDIAN EXPERIENCER CONSTRUCTION AND THE ENGLISH SEEM CONSTRUCTION K. A. Jayaseelan CIEFL, Hyderabad 1. Introduction In many languages e.g. Malayalam, Tamil, Hindi, the same verb is used in the Experiencer
More informationParsing with CFGs L445 / L545 / B659. Dept. of Linguistics, Indiana University Spring Parsing with CFGs. Direction of processing
L445 / L545 / B659 Dept. of Linguistics, Indiana University Spring 2016 1 / 46 : Overview Input: a string Output: a (single) parse tree A useful step in the process of obtaining meaning We can view the
More informationParsing with CFGs. Direction of processing. Top-down. Bottom-up. Left-corner parsing. Chart parsing CYK. Earley 1 / 46.
: Overview L545 Dept. of Linguistics, Indiana University Spring 2013 Input: a string Output: a (single) parse tree A useful step in the process of obtaining meaning We can view the problem as searching
More informationLECTURER: BURCU CAN Spring
LECTURER: BURCU CAN 2017-2018 Spring Regular Language Hidden Markov Model (HMM) Context Free Language Context Sensitive Language Probabilistic Context Free Grammar (PCFG) Unrestricted Language PCFGs can
More informationNatural Language Processing CS Lecture 06. Razvan C. Bunescu School of Electrical Engineering and Computer Science
Natural Language Processing CS 6840 Lecture 06 Razvan C. Bunescu School of Electrical Engineering and Computer Science bunescu@ohio.edu Statistical Parsing Define a probabilistic model of syntax P(T S):
More informationA* Search. 1 Dijkstra Shortest Path
A* Search Consider the eight puzzle. There are eight tiles numbered 1 through 8 on a 3 by three grid with nine locations so that one location is left empty. We can move by sliding a tile adjacent to the
More informationCAS LX 522 Syntax I November 4, 2002 Week 9: Wh-movement, supplement
CAS LX 522 Syntax I November 4, 2002 Fall 2002 Week 9: Wh-movement, supplement () Italian Tuo fratello ( your brother ), [ CP a cui i [ TP mi domando [ CP che storie i [ TP abbiano raccontato t i t j...
More informationIntroduction to Semantics. The Formalization of Meaning 1
The Formalization of Meaning 1 1. Obtaining a System That Derives Truth Conditions (1) The Goal of Our Enterprise To develop a system that, for every sentence S of English, derives the truth-conditions
More informationWh-movement. CAS LX 522 Syntax I Fall 2001 November 6, 2001
AS LX 522 Syntax I Fall 200 November 6, 200 Paul Hagstrom Week 9: Wh-movement Preliminary tree to remind ourselves: They will bake a cake. () P [ Q] T T VP will P bake they NP a N cake Verb bake needs
More informationParsing. Based on presentations from Chris Manning s course on Statistical Parsing (Stanford)
Parsing Based on presentations from Chris Manning s course on Statistical Parsing (Stanford) S N VP V NP D N John hit the ball Levels of analysis Level Morphology/Lexical POS (morpho-synactic), WSD Elements
More informationDependency grammar. Recurrent neural networks. Transition-based neural parsing. Word representations. Informs Models
Dependency grammar Morphology Word order Transition-based neural parsing Word representations Recurrent neural networks Informs Models Dependency grammar Morphology Word order Transition-based neural parsing
More informationConstituency. Doug Arnold
Constituency Doug Arnold doug@essex.ac.uk Spose we have a string... xyz..., how can we establish whether xyz is a constituent (i.e. syntactic unit); i.e. whether the representation of... xyz... should
More informationLing 98a: The Meaning of Negation (Week 5)
Yimei Xiang yxiang@fas.harvard.edu 15 October 2013 1 Review Negation in propositional logic, oppositions, term logic of Aristotle Presuppositions Projection and accommodation Three-valued logic External/internal
More informationUnit 2: Tree Models. CS 562: Empirical Methods in Natural Language Processing. Lectures 19-23: Context-Free Grammars and Parsing
CS 562: Empirical Methods in Natural Language Processing Unit 2: Tree Models Lectures 19-23: Context-Free Grammars and Parsing Oct-Nov 2009 Liang Huang (lhuang@isi.edu) Big Picture we have already covered...
More informationModels of Adjunction in Minimalist Grammars
Models of Adjunction in Minimalist Grammars Thomas Graf mail@thomasgraf.net http://thomasgraf.net Stony Brook University FG 2014 August 17, 2014 The Theory-Neutral CliffsNotes Insights Several properties
More informationHPSG II: the plot thickens
Syntactic Models 2/21/06 HPSG II: the plot thickens 1 Passive: a lexical rule that rearranges ARG-ST! (1) Passive Lexical Rule < 1, tv - lxm ARG - ST INDEX i < FPSP 1 a, > part - lxm SYN HEAD FORM pass
More informationFeatures. An argument DP must have been assigned Case by S-structure. A Specifier of IP must have been occupied by something by S-structure.
Features 26 October 2015 We left with a system that included the following elements. (1) Head Movement onstraint IfanX 0 movestoay 0,therecanbenoZ 0 thatisc-commandedbyy 0 and c-commandsx 0. (2) Like(s)
More informationSome binding facts. Binding in HPSG. The three basic principles. Binding theory of Chomsky
Some binding facts Binding in HPSG Binding theory has to do with constraining referentially dependent elements in terms of what is a possible antecedent, or binder. Introduction to HPSG June, 009 (1) John
More informationRecap: Lexicalized PCFGs (Fall 2007): Lecture 5 Parsing and Syntax III. Recap: Charniak s Model. Recap: Adding Head Words/Tags to Trees
Recap: Lexicalized PCFGs We now need to estimate rule probabilities such as P rob(s(questioned,vt) NP(lawyer,NN) VP(questioned,Vt) S(questioned,Vt)) 6.864 (Fall 2007): Lecture 5 Parsing and Syntax III
More informationParsing with Context-Free Grammars
Parsing with Context-Free Grammars CS 585, Fall 2017 Introduction to Natural Language Processing http://people.cs.umass.edu/~brenocon/inlp2017 Brendan O Connor College of Information and Computer Sciences
More informationCS 188 Introduction to AI Fall 2005 Stuart Russell Final
NAME: SID#: Section: 1 CS 188 Introduction to AI all 2005 Stuart Russell inal You have 2 hours and 50 minutes. he exam is open-book, open-notes. 100 points total. Panic not. Mark your answers ON HE EXAM
More informationHPSG: Binding Theory
HPSG: Binding Theory Doug Arnold doug@essexacuk Introduction Binding Theory is to do with the syntactic restrictions on the distribution of referentially dependent items and their antecedents: reflexives/reciprocals
More informationCS460/626 : Natural Language
CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 23, 24 Parsing Algorithms; Parsing in case of Ambiguity; Probabilistic Parsing) Pushpak Bhattacharyya CSE Dept., IIT Bombay 8 th,
More informationA Context-Free Grammar
Statistical Parsing A Context-Free Grammar S VP VP Vi VP Vt VP VP PP DT NN PP PP P Vi sleeps Vt saw NN man NN dog NN telescope DT the IN with IN in Ambiguity A sentence of reasonable length can easily
More informationNatural Language Processing : Probabilistic Context Free Grammars. Updated 5/09
Natural Language Processing : Probabilistic Context Free Grammars Updated 5/09 Motivation N-gram models and HMM Tagging only allowed us to process sentences linearly. However, even simple sentences require
More informationRecap: Tree geometry, selection, Θ-theory
Syntax II Seminar 2 Recap: Tree geometry, selection, Θ-theory Dr. James Grifiihs james.grifiihs@uni-konsianz.de Recap: tree geometry terminology M N O D E F H I J Branches 2 Recap: tree geometry terminology
More informationThe Formal Architecture of. Lexical-Functional Grammar. Ronald M. Kaplan and Mary Dalrymple
The Formal Architecture of Lexical-Functional Grammar Ronald M. Kaplan and Mary Dalrymple Xerox Palo Alto Research Center 1. Kaplan and Dalrymple, ESSLLI 1995, Barcelona Architectural Issues Representation:
More informationGrundlagenmodul Semantik All Exercises
Grundlagenmodul Semantik All Exercises Sommersemester 2014 Exercise 1 Are the following statements correct? Justify your answers in a single short sentence. 1. 11 {x x is a square number} 2. 11 {x {y y
More informationAn introduction to German Syntax. 1. Head directionality: A major source of linguistic divergence
An introduction to German Syntax 19 January 2018 1. Head directionality: A major source of linguistic divergence In English, heads uniformly precede their complements: (1) a. [ kiss Mary] a. * [ Mary kiss]
More informationClass Notes: Tsujimura (2007), Ch. 5. Syntax (1), pp (3) a. [[akai hon]-no hyooshi] b. [akai [hon-no hyooshi]]
Class otes: Tsujimura (2007), Ch. 5. yntax (1), pp. 206-220 p. 206 What is!ytx"?! n area in linguistics that deals with the REGULRLITY of how words are put together to create grammatical sentences What
More informationHandout 8: Computation & Hierarchical parsing II. Compute initial state set S 0 Compute initial state set S 0
Massachusetts Institute of Technology 6.863J/9.611J, Natural Language Processing, Spring, 2001 Department of Electrical Engineering and Computer Science Department of Brain and Cognitive Sciences Handout
More informationSharpening the empirical claims of generative syntax through formalization
Sharpening the empirical claims of generative syntax through formalization Tim Hunter University of Minnesota, Twin Cities ESSLLI, August 2015 Part 1: Grammars and cognitive hypotheses What is a grammar?
More informationTime Zones - KET Grammar
Inventory of grammatical areas Verbs Regular and irregular forms Pages 104-105 (Unit 1 Modals can (ability; requests; permission) could (ability; polite requests) Page 3 (Getting Started) Pages 45-47,
More information2013 ISSN: JATLaC Journal 8: t 1. t t Chomsky 1993 I Radford (2009) R I t t R I 2. t R t (1) (= R's (15), p. 86) He could have helped
t 1. tt Chomsky 1993 IRadford (2009) R Itt R I 2. t R t (1) (= R's (15), p. 86) He could have helped her, or [she have helped him]. 2 have I has had I I could ellipsis I gapping R (2) (= R (18), p.88)
More informationStepanov 2007: The End of CED? Minimalism and Extraction Domains
Stepanov (2007) Stepanov 2007: The End of CED? Minimalism and Extraction Domains 1 Introduction In English (and other languages), overt wh-extraction out of subjects or adjuncts (as opposed to objects)
More informationCS626: NLP, Speech and the Web. Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 14: Parsing Algorithms 30 th August, 2012
CS626: NLP, Speech and the Web Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 14: Parsing Algorithms 30 th August, 2012 Parsing Problem Semantics Part of Speech Tagging NLP Trinity Morph Analysis
More informationIntroduction to Semantics. Common Nouns and Adjectives in Predicate Position 1
Common Nouns and Adjectives in Predicate Position 1 (1) The Lexicon of Our System at Present a. Proper Names: [[ Barack ]] = Barack b. Intransitive Verbs: [[ smokes ]] = [ λx : x D e. IF x smokes THEN
More informationS NP VP 0.9 S VP 0.1 VP V NP 0.5 VP V 0.1 VP V PP 0.1 NP NP NP 0.1 NP NP PP 0.2 NP N 0.7 PP P NP 1.0 VP NP PP 1.0. N people 0.
/6/7 CS 6/CS: Natural Language Processing Instructor: Prof. Lu Wang College of Computer and Information Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang The grammar: Binary, no epsilons,.9..5
More informationSemantics and Generative Grammar. Quantificational DPs, Part 3: Covert Movement vs. Type Shifting 1
Quantificational DPs, Part 3: Covert Movement vs. Type Shifting 1 1. Introduction Thus far, we ve considered two competing analyses of sentences like those in (1). (1) Sentences Where a Quantificational
More informationUniformity of Theta- Assignment Hypothesis. Transit i ve s. Unaccusatives. Unergatives. Double object constructions. Previously... Bill lied.
Preiously... Uniformity of Theta- Assignment Hypothesis, daughter of P = Agent, daughter of P = Theme PP, daughter of! = Goal P called P Chris P P books gae P to PP Chris Unaccusaties Transit i e s The
More informationCMPT-825 Natural Language Processing. Why are parsing algorithms important?
CMPT-825 Natural Language Processing Anoop Sarkar http://www.cs.sfu.ca/ anoop October 26, 2010 1/34 Why are parsing algorithms important? A linguistic theory is implemented in a formal system to generate
More informationIntellectual Property of Mariamalia Hidalgo
Grammar Review Present Tense of BE Present I am ( m) am not you are ( re) aren t Affirmative sentences Subject + verb be + object. he is ( s) isn t sentences Subject + (verb be + not) + object. she is
More informationContext-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing
Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing Natural Language Processing! CS 6120 Spring 2014! Northeastern University!! David Smith! with some slides from Jason Eisner & Andrew
More informationHierarchical and Linear Constraints on Structure
Hierarchical and Linear Constraints on Structure Peter Sells Sandbjerg Ph.D. course June 14 16, 2006 Part II 1 OT Evaluations (1) CP XP C C IP Pro/Adv NP Adv I I Adv/Neg/NQ/Pro VP V Pro V NP NP PP V Prt
More informationProseminar on Semantic Theory Fall 2013 Ling 720 The Proper Treatment of Quantification in Ordinary English, Part 1: The Fragment of English
The Proper Treatment of Quantification in Ordinary English, Part 1: The Fragment of English We will now explore the analysis of English that Montague puts forth in his seminal paper, PTQ. As we ve already
More informationCategories and Transformations 321
Categories and ransformations 321 4.10 Functional Categories and Formal Features What precedes substantially revises the framework developed in chapters 1 3. But we have not yet subjected functional categories
More informationCS 662 Sample Midterm
Name: 1 True/False, plus corrections CS 662 Sample Midterm 35 pts, 5 pts each Each of the following statements is either true or false. If it is true, mark it true. If it is false, correct the statement
More informationCS 6120/CS4120: Natural Language Processing
CS 6120/CS4120: Natural Language Processing Instructor: Prof. Lu Wang College of Computer and Information Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Assignment/report submission
More informationIntroduction to Semantics. Pronouns and Variable Assignments. We ve seen that implicatures are crucially related to context.
Pronouns and Variable Assignments 1. Putting this Unit in Context (1) What We ve Done So Far This Unit Expanded our semantic theory so that it includes (the beginnings of) a theory of how the presuppositions
More informationUnterspezifikation in der Semantik Scope Semantics in Lexicalized Tree Adjoining Grammars
in der emantik cope emantics in Lexicalized Tree Adjoining Grammars Laura Heinrich-Heine-Universität Düsseldorf Wintersemester 2011/2012 LTAG: The Formalism (1) Tree Adjoining Grammars (TAG): Tree-rewriting
More informationSemantics and Generative Grammar. Pronouns and Variable Assignments 1. We ve seen that implicatures are crucially related to context.
Pronouns and Variable Assignments 1 1. Putting this Unit in Context (1) What We ve Done So Far This Unit Expanded our semantic theory so that it includes (the beginnings of) a theory of how the presuppositions
More informationSpring 2017 Ling 620. An Introduction to the Semantics of Tense 1
1. Introducing Evaluation Times An Introduction to the Semantics of Tense 1 (1) Obvious, Fundamental Fact about Sentences of English The truth of some sentences (of English) depends upon the time they
More informationSpring 2018 Ling 620 Introduction to Semantics of Questions: Questions as Sets of Propositions (Hamblin 1973, Karttunen 1977)
Introduction to Semantics of Questions: Questions as Sets of Propositions (Hamblin 1973, Karttunen 1977) 1. Question Meanings and Sets of Propositions (1) The Semantics of Declarative Sentence Dave smokes
More informationComputational Linguistics
Computational Linguistics Dependency-based Parsing Clayton Greenberg Stefan Thater FR 4.7 Allgemeine Linguistik (Computerlinguistik) Universität des Saarlandes Summer 2016 Acknowledgements These slides
More informationComputational Linguistics. Acknowledgements. Phrase-Structure Trees. Dependency-based Parsing
Computational Linguistics Dependency-based Parsing Dietrich Klakow & Stefan Thater FR 4.7 Allgemeine Linguistik (Computerlinguistik) Universität des Saarlandes Summer 2013 Acknowledgements These slides
More informationContext-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing
Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing Natural Language Processing CS 4120/6120 Spring 2017 Northeastern University David Smith with some slides from Jason Eisner & Andrew
More informationSimpler Syntax. Ling : Sign-Based Construction Grammar Instructor: Ivan A. Sag URL:
Simpler Syntax Ling 7800-065: Sign-Based Construction Grammar Instructor: Ivan A. Sag (sag@stanford.edu) URL: http://lingo.stanford.edu/sag/li11-sbcg 1/ 58 Constructs ] [mtr sign construct : dtrs nelist(sign)
More informationBARE PHRASE STRUCTURE, LCA AND LINEARIZATION
Neuere Entwicklungen der Grammatiktheorie Universität Wien, SoSe 2008 Winfried Lechner wlechner@gs.uoa.gr Handout #2 BARE PHRASE STRUCTURE, LCA AND LINEARIZATION http://vivaldi.sfs.nphil.uni-tuebingen.de/%7ennsle01/synvie.htm
More informationControl and Tough- Movement
Control and Tough- Movement Carl Pollard February 2, 2012 Control (1/5) We saw that PRO is used for the unrealized subject of nonfinite verbals and predicatives where the subject plays a semantic role
More informationStructures mathématiques du langage
tructures mathématiques du langage Alain Lecomte 16 février 2014 1 Heim and Kratzer s theory Montague s grammar was conceived and built during the sixties of the last century, without much attention paid
More informationDT2118 Speech and Speaker Recognition
DT2118 Speech and Speaker Recognition Language Modelling Giampiero Salvi KTH/CSC/TMH giampi@kth.se VT 2015 1 / 56 Outline Introduction Formal Language Theory Stochastic Language Models (SLM) N-gram Language
More informationCS 224N HW:#3. (V N0 )δ N r p r + N 0. N r (r δ) + (V N 0)δ. N r r δ. + (V N 0)δ N = 1. 1 we must have the restriction: δ NN 0.
CS 224 HW:#3 ARIA HAGHIGHI SUID :# 05041774 1. Smoothing Probability Models (a). Let r be the number of words with r counts and p r be the probability for a word with r counts in the Absolute discounting
More informationComputer Sciences Department
Computer Sciences Department 1 Reference Book: INTRODUCTION TO THE THEORY OF COMPUTATION, SECOND EDITION, by: MICHAEL SIPSER Computer Sciences Department 3 ADVANCED TOPICS IN C O M P U T A B I L I T Y
More informationChiastic Lambda-Calculi
Chiastic Lambda-Calculi wren ng thornton Cognitive Science & Computational Linguistics Indiana University, Bloomington NLCS, 28 June 2013 wren ng thornton (Indiana University) Chiastic Lambda-Calculi NLCS,
More informationConstituency Parsing
CS5740: Natural Language Processing Spring 2017 Constituency Parsing Instructor: Yoav Artzi Slides adapted from Dan Klein, Dan Jurafsky, Chris Manning, Michael Collins, Luke Zettlemoyer, Yejin Choi, and
More informationModel-Theory of Property Grammars with Features
Model-Theory of Property Grammars with Features Denys Duchier Thi-Bich-Hanh Dao firstname.lastname@univ-orleans.fr Yannick Parmentier Abstract In this paper, we present a model-theoretic description of
More informationTwo-phase Implementation of Morphological Analysis
Two-phase Implementation of Morphological Analysis Arvi Hurskainen Institute for Asian and African Studies, Box 59 FIN-00014 University of Helsinki, Finland arvi.hurskainen@helsinki.fi Abstract SALAMA
More information564 Lecture 25 Nov. 23, Continuing note on presuppositional vs. nonpresuppositional dets.
564 Lecture 25 Nov. 23, 1999 1 Continuing note on presuppositional vs. nonpresuppositional dets. Here's the argument about the nonpresupp vs. presupp analysis of "every" that I couldn't reconstruct last
More informationControl and Tough- Movement
Department of Linguistics Ohio State University February 2, 2012 Control (1/5) We saw that PRO is used for the unrealized subject of nonfinite verbals and predicatives where the subject plays a semantic
More informationSpring 2018 Ling 620 The Basics of Intensional Semantics, Part 1: The Motivation for Intensions and How to Formalize Them 1
The Basics of Intensional Semantics, Part 1: The Motivation for Intensions and How to Formalize Them 1 1. The Inadequacies of a Purely Extensional Semantics (1) Extensional Semantics a. The interpretation
More informationLing 5801: Lecture Notes 7 From Programs to Context-Free Grammars
Ling 5801: Lecture otes 7 From Programs to Context-Free rammars 1. The rules we used to define programs make up a context-free grammar A Context-Free rammar is a tuple C,X,S,R, where: C is a finite set
More informationPOS-Tagging. Fabian M. Suchanek
POS-Tagging Fabian M. Suchanek 100 Def: POS A Part-of-Speech (also: POS, POS-tag, word class, lexical class, lexical category) is a set of words with the same grammatical role. Alizée wrote a really great
More informationDialogue Systems. Statistical NLU component. Representation. A Probabilistic Dialogue System. Task: map a sentence + context to a database query
Statistical NLU component Task: map a sentence + context to a database query Dialogue Systems User: Show me flights from NY to Boston, leaving tomorrow System: [returns a list of flights] Origin (City
More informationContext- Free Parsing with CKY. October 16, 2014
Context- Free Parsing with CKY October 16, 2014 Lecture Plan Parsing as Logical DeducBon Defining the CFG recognibon problem BoHom up vs. top down Quick review of Chomsky normal form The CKY algorithm
More informationOne hint from secondary predication (from Baker 1997 (8) A secondary predicate cannot take the goal argument as subject of predication, wheth
MIT, Fall 2003 1 The Double Object Construction (Larson 1988, Aoun & Li 1989) MIT, 24.951, Fr 14 Nov 2003 A familiar puzzle The Dative Alternation (1) a. I gave the candy to the children b. I gave the
More informationNatural Language Processing
SFU NatLangLab Natural Language Processing Anoop Sarkar anoopsarkar.github.io/nlp-class Simon Fraser University September 27, 2018 0 Natural Language Processing Anoop Sarkar anoopsarkar.github.io/nlp-class
More informationProbabilistic Context-free Grammars
Probabilistic Context-free Grammars Computational Linguistics Alexander Koller 24 November 2017 The CKY Recognizer S NP VP NP Det N VP V NP V ate NP John Det a N sandwich i = 1 2 3 4 k = 2 3 4 5 S NP John
More informationLogical Translations Jean Mark Gawron San Diego State University. 1 Introduction 2
Logical Translations Jean Mark Gawron San Diego State University Contents 1 Introduction 2 2 Truth-Functional Connectives 2 2.1 And................................ 2 2.2 Or.................................
More information10/17/04. Today s Main Points
Part-of-speech Tagging & Hidden Markov Model Intro Lecture #10 Introduction to Natural Language Processing CMPSCI 585, Fall 2004 University of Massachusetts Amherst Andrew McCallum Today s Main Points
More informationSyntax and Semantics in Minimalist Grammars
Syntax and Semantics in Minimalist Grammars Gregory M. Kobele ESSLLI 09 This chapter presents a grammar for a fragment of English A-movement. 1 The constructions accounted for include raising, passivization,
More informationHandout 3: PTQ Revisited (Muskens 1995, Ch. 4)
Handout 3: PTQ Revisited (Muskens 1995, Ch. 4) Semantics C (Spring 2010) Montague s PTQ article (Montague 1973), the paper in which he gave his Proper Treatment of Quantification, functions as the paradigm
More informationGrammar and Feature Unification
Grammar and Feature Unification Problems with CF Phrase Structure Grammars Difficult to capture dependencies between constituents the boy runs the boys run * the boy run * the boys runs Problems with CF
More informationTree Adjoining Grammars
Tree Adjoining Grammars Feature Structure Based TAG Laura Kallmeyer & Benjamin Burkhardt HHU Düsseldorf WS 2017/2018 1 / 20 Outline 1 Why feature structures? 2 Basics of feature structure logic 3 Feature
More informationTwo Reconstruction Puzzles Yael Sharvit University of Connecticut
Workshop on Direct Compositionality June 19-21, 2003 Brown University Two Reconstruction Puzzles Yael Sharvit University of Connecticut yael.sharvit@uconn.edu Some constructions exhibit what is known as
More informationParasitic Scope (Barker 2007) Semantics Seminar 11/10/08
Parasitic Scope (Barker 2007) Semantics Seminar 11/10/08 1. Overview Attempts to provide a compositional, fully semantic account of same. Elements other than NPs in particular, adjectives can be scope-taking
More informationReview. Earley Algorithm Chapter Left Recursion. Left-Recursion. Rule Ordering. Rule Ordering
Review Earley Algorithm Chapter 13.4 Lecture #9 October 2009 Top-Down vs. Bottom-Up Parsers Both generate too many useless trees Combine the two to avoid over-generation: Top-Down Parsing with Bottom-Up
More information