Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically

Size: px
Start display at page:

Download "Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically"

Transcription

1 Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically NP S AUX VP Ali will V NP help D N the man A node is any point in the tree diagram and it can be: Branching node like S and lower NP. Non-branching node like AUX and V. Terminal nodes: lexical items at end of tree like help Nodes are related to each other by two relations:

2 1. Dominance: A node X dominates node Y if: X is higher than Y and X is connected to Y by a branch. For example, NP dominates Ali, VP dominates D and NP. Immediate dominance: A node immediately dominates another if there s s no intervening node. e.g. S immediately dominates NP, AUX, VP but not help. 2. Precedence: A node X precedes Y if it is on the left and both aren t dominating each other. e.g. Ali precedes will,, but the doesn t precede man.

3 Phrase Structure Rules These rules derive different types of phrases and unlimited number of sentences: NP (D) (AdjP( AdjP) ) N This rule cannot predict the structure of the phrase and we need to memorize infinite number of rules. The Structure of Phrases 1. VP: It consists of - lexical category: head V - Phrasal category or maximal projection: : VP as a whole - Intermediate category: V V (part of VP)

4 V V is a level that contains the object and verb s s modifiers: The head verb (and its object if there is one) is in the lowest V level The verb s s modifiers are placed in higher V V levels, called adjunct Subject combines with highest V V level, i.e. specifier of VP For example: they eat lunch in school tonight VP spec V V they V V NP V PP tonight V NP in school eat lunch

5 Evidence for V V V is a constituent structure that can be replaced by do so: - They eat lunch in school tonight and Ali does so. (eat lunch in school tonight) - They eat lunch in school tonight and Ali did so this morning. (eat lunch in school) - They eat lunch in school tonight and Ali does so at work this morning. (eat lunch) adjuncts are recursive, i.e. repeatedly added. 2. NP N is a level that contains the object and noun s s modifiers: The head noun (and its object if there is one) is in the lowest N N level The noun s s modifiers are placed in higher N N levels, called adjunct The determiner combines with the highest N, N, i.e. specifier of VP For example, the big book of poems with the blue cover

6 NP Det N 1 the N 2N AP N 3 N 3 PP big N PP with blue cover book of poems Evidence for N N N is a constituent structure that can be replaced by one: - I want this [big book of poems with the blue cover] not that one (N 1) - I want this big [book of poems with the blue cover] not that small one (N 2) - I want this big [book of poems] with the blue cover not that small one with the red cover (N 3) 3. AdjP & PP We apply the same structure to these phrases:

7 He is [quite jealous of Ali]. He stood [right across the bridge] The lowest P P and adj includes these heads (and their complements). The spec combines with P P and adj and is placed by modifiers (like quite, very, rather, so for the adj and straight, right for the prep). General structure (x-bar) We can have these general rules that exactly predict the structure re of different types of phrases: XP spec X X (specifier) X X YP (head and complement) X X YP (adjunct= modifier)

8 Sentence structure Aux is the head of the sentence because it carries tense and agreement: They are working hard. But what about non-auxiliary verbs: They worked hard Aux is the head even though it isn t t overt. Evidence for aux as head of sentence: 1. Cleft sentence: work hard, they did indeed. The tense is on aux and not part of VP.

9 2. Pseudo-cleft: what they did was work hard. Tense is part of a node, Infl(ection),, which can be filled with overt aux or left empty. s NP I VP +tense +agr I is finite because it has (+t, +agr+ agr). Infinitive clause: I ask [Ali to work hard] To is the head of the infin clause and it s -t, -agr, non-finite I

10 The structure of IP IP NP I I VP will V V -ed V NP to finish the work IP spec I I I I VP VP is always a complement of I.

11 IP is a functional category, not a lexical category, because it is used for grammatical function: t, agr Complementizer phrase (CP) CP is another functional category since its head, c, introduces a subordinate clause: C= that, for [-wh[ wh] if, whether [+wh wh] I believe [that Ali will work hard]. I want [for Ali to work hard]. I wonder [if/whether Ali worked hard].

12 The above CP clauses have the following structure: CP spec C C C IP Ali will worked hard - IP is always the complement of C. - C is filled by that, for and moved will,, forming yes/no question: CP spec C C C IP will NP I [+wh wh] ] Ali I VP work hard

13 - C can be filled by either will or that, not both - Spec is filled by whether, if and wh-questions: Structure of CP: CP Spec C C C C IP CP spec C C when C IP will NP I Ali I VP work hard

14 Structural relations Government: A head governs its phrase e.g. N heads its phrase, I the IP, and C the CP. Agreement between head and non-head is established under government: 1. NP agreement NP Det N this/these N PP car/cars of Ali

15 There s s agreement in number between head (N) and spec (det). But agreement is poor in English since it s s shown morphologically in number not in gender and person as in French. 2. sub/ verb agreement There spec-head agreement between Infl and spec (sub) in number, person: IP spec I I He/they I VP -s s / ø play

16 3. CP agreement There s s a spec-head agreement in CP in [wh[ wh] ] feature: I wonder if he played. CP spec C C if C IP [+wh wh] ] [+wh wh] ] played C-command A c-commands c commands B if and if i. A doesn t t dominate B and B doesn t t dominate A; and ii. The first branching node dominating A dominates B. For instance, in (2) the spec c-commands commands every node in IP, but VP only I.

17 Let us now refine the notion of government: Government: A governs B if and if: A is a governor (i.e. a head) and A c-commands c commands B. Strict c-command c command & M-commandM 1. Strict c-commandc command VP V V PP V NP P P NP ate the food in the garden V strictly c-commands c commands NP because they re dominated by 1 st branching node but not PP.

18 2. M-commandM The head c-commands c commands the adjunct in the garden since there is a node VP that dominate both of them: A c-commands c commands B if and only if: A doesn t t dominate B and every node X (maximal projection) dominating A dominates B. V c-commands c commands NP, V doesn t t c-commands c commands PP (dominance by1 st branching node) V m-commands m NP, PP, NP (inside PP) (dominance by maximal proj.) P doesn t t m-commands m V (PP doesn t t dominate V)

19 Government is refined as: A governs B if and if: i. A is a governor; and ii. A m-commands m B; and iii. No barrier between A and B. In (1), the verb ate governs the PP, but it doesn t t govern the NP since it is a barrier. However the verb m-commands m NP.

X-bar theory. X-bar :

X-bar theory. X-bar : is one of the greatest contributions of generative school in the filed of knowledge system. Besides linguistics, computer science is greatly indebted to Chomsky to have propounded the theory of x-bar.

More information

Parsing with Context-Free Grammars

Parsing with Context-Free Grammars Parsing with Context-Free Grammars Berlin Chen 2005 References: 1. Natural Language Understanding, chapter 3 (3.1~3.4, 3.6) 2. Speech and Language Processing, chapters 9, 10 NLP-Berlin Chen 1 Grammars

More information

(7) a. [ PP to John], Mary gave the book t [PP]. b. [ VP fix the car], I wonder whether she will t [VP].

(7) a. [ PP to John], Mary gave the book t [PP]. b. [ VP fix the car], I wonder whether she will t [VP]. CAS LX 522 Syntax I Fall 2000 September 18, 2000 Paul Hagstrom Week 2: Movement Movement Last time, we talked about subcategorization. (1) a. I can solve this problem. b. This problem, I can solve. (2)

More information

CAS LX 522 Syntax I Fall 2000 October 10, 2000 Week 5: Case Theory and θ Theory. θ-theory continued

CAS LX 522 Syntax I Fall 2000 October 10, 2000 Week 5: Case Theory and θ Theory. θ-theory continued CAS LX 522 Syntax I Fall 2000 October 0, 2000 Paul Hagstrom Week 5: Case Theory and θ Theory θ-theory continued From last time: verbs have θ-roles (e.g., Agent, Theme, ) to assign, specified in the lexicon

More information

Another look at PSRs: Intermediate Structure. Starting X-bar theory

Another look at PSRs: Intermediate Structure. Starting X-bar theory Another look at PSRs: Intermediate Structure Starting X-bar theory Andrew Carnie, 2006 Substitution Andrew Carnie, 2006 Substitution If a group of words can be replaced by a single word, they are a constituent.

More information

Computationele grammatica

Computationele grammatica Computationele grammatica Docent: Paola Monachesi Contents First Last Prev Next Contents 1 Unbounded dependency constructions (UDCs)............... 3 2 Some data...............................................

More information

1. Background. Task: Determine whether a given string of words is a grammatical (well-formed) sentence of language L i or not.

1. Background. Task: Determine whether a given string of words is a grammatical (well-formed) sentence of language L i or not. Constraints in Syntax [1] Phrase Structure and Derivations Düsseldorf LSA/DGfS Summerschool 2002 Gereon Müller (IDS Mannheim) gereon.mueller@ids-mannheim.de 1. Background Task: Determine whether a given

More information

Other types of Movement

Other types of Movement Other types of Movement So far we seen Wh-movement, which moves certain types of (XP) constituents to the specifier of a CP. Wh-movement is also called A-bar movement. We will look at two more types of

More information

Ling 240 Lecture #15. Syntax 4

Ling 240 Lecture #15. Syntax 4 Ling 240 Lecture #15 Syntax 4 agenda for today Give me homework 3! Language presentation! return Quiz 2 brief review of friday More on transformation Homework 4 A set of Phrase Structure Rules S -> (aux)

More information

Artificial Intelligence

Artificial Intelligence CS344: Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 20-21 Natural Language Parsing Parsing of Sentences Are sentences flat linear structures? Why tree? Is

More information

CS 712: Topics in NLP Linguistic Phrases and Statistical Phrases

CS 712: Topics in NLP Linguistic Phrases and Statistical Phrases CS 712: Topics in NLP Linguistic Phrases and Statistical Phrases Pushpak Bhattacharyya, CSE Department, IIT Bombay 18 March, 2013 (main text: Syntax by Adrew Carnie, Blackwell Publication, 2002) Domination

More information

CAS LX 523 Syntax II Spring 2001 March 13, (1) A qp. Kayne, Richard (1995). The antisymmetry of syntax. Cambridge, MA: MIT Press.

CAS LX 523 Syntax II Spring 2001 March 13, (1) A qp. Kayne, Richard (1995). The antisymmetry of syntax. Cambridge, MA: MIT Press. CAS LX 52 Syntax II Spring 200 March, 200 Paul Hagstrom Week 7: Antisymmetry Kayne, Richard (995). The antisymmetry of syntax. Cambridge, MA: MIT Press. Koopman, Hilda (2000). The spec-head configuration.

More information

2 A not-quite-argument for X-bar structure in noun phrases

2 A not-quite-argument for X-bar structure in noun phrases CAS LX 321 / GRS LX 621 Syntax: Introduction to Sentential Structure ovember 16, 2017 1 and pronouns (1) he linguists yodel. (2) We linguists yodel. (3) hey looked at us linguists. (4) hey looked at linguists.

More information

Andrew Carnie, Structural Relations. The mathematical properties of phrase structure trees

Andrew Carnie, Structural Relations. The mathematical properties of phrase structure trees Structural Relations The mathematical properties of phrase structure trees Important! Important! Even if you have trouble with the formal definitions, try to understand the INTUITIVE idea behind them.

More information

THE DRAVIDIAN EXPERIENCER CONSTRUCTION AND THE ENGLISH SEEM CONSTRUCTION. K. A. Jayaseelan CIEFL, Hyderabad

THE DRAVIDIAN EXPERIENCER CONSTRUCTION AND THE ENGLISH SEEM CONSTRUCTION. K. A. Jayaseelan CIEFL, Hyderabad THE DRAVIDIAN EXPERIENCER CONSTRUCTION AND THE ENGLISH SEEM CONSTRUCTION K. A. Jayaseelan CIEFL, Hyderabad 1. Introduction In many languages e.g. Malayalam, Tamil, Hindi, the same verb is used in the Experiencer

More information

Parsing with CFGs L445 / L545 / B659. Dept. of Linguistics, Indiana University Spring Parsing with CFGs. Direction of processing

Parsing with CFGs L445 / L545 / B659. Dept. of Linguistics, Indiana University Spring Parsing with CFGs. Direction of processing L445 / L545 / B659 Dept. of Linguistics, Indiana University Spring 2016 1 / 46 : Overview Input: a string Output: a (single) parse tree A useful step in the process of obtaining meaning We can view the

More information

Parsing with CFGs. Direction of processing. Top-down. Bottom-up. Left-corner parsing. Chart parsing CYK. Earley 1 / 46.

Parsing with CFGs. Direction of processing. Top-down. Bottom-up. Left-corner parsing. Chart parsing CYK. Earley 1 / 46. : Overview L545 Dept. of Linguistics, Indiana University Spring 2013 Input: a string Output: a (single) parse tree A useful step in the process of obtaining meaning We can view the problem as searching

More information

LECTURER: BURCU CAN Spring

LECTURER: BURCU CAN Spring LECTURER: BURCU CAN 2017-2018 Spring Regular Language Hidden Markov Model (HMM) Context Free Language Context Sensitive Language Probabilistic Context Free Grammar (PCFG) Unrestricted Language PCFGs can

More information

Natural Language Processing CS Lecture 06. Razvan C. Bunescu School of Electrical Engineering and Computer Science

Natural Language Processing CS Lecture 06. Razvan C. Bunescu School of Electrical Engineering and Computer Science Natural Language Processing CS 6840 Lecture 06 Razvan C. Bunescu School of Electrical Engineering and Computer Science bunescu@ohio.edu Statistical Parsing Define a probabilistic model of syntax P(T S):

More information

A* Search. 1 Dijkstra Shortest Path

A* Search. 1 Dijkstra Shortest Path A* Search Consider the eight puzzle. There are eight tiles numbered 1 through 8 on a 3 by three grid with nine locations so that one location is left empty. We can move by sliding a tile adjacent to the

More information

CAS LX 522 Syntax I November 4, 2002 Week 9: Wh-movement, supplement

CAS LX 522 Syntax I November 4, 2002 Week 9: Wh-movement, supplement CAS LX 522 Syntax I November 4, 2002 Fall 2002 Week 9: Wh-movement, supplement () Italian Tuo fratello ( your brother ), [ CP a cui i [ TP mi domando [ CP che storie i [ TP abbiano raccontato t i t j...

More information

Introduction to Semantics. The Formalization of Meaning 1

Introduction to Semantics. The Formalization of Meaning 1 The Formalization of Meaning 1 1. Obtaining a System That Derives Truth Conditions (1) The Goal of Our Enterprise To develop a system that, for every sentence S of English, derives the truth-conditions

More information

Wh-movement. CAS LX 522 Syntax I Fall 2001 November 6, 2001

Wh-movement. CAS LX 522 Syntax I Fall 2001 November 6, 2001 AS LX 522 Syntax I Fall 200 November 6, 200 Paul Hagstrom Week 9: Wh-movement Preliminary tree to remind ourselves: They will bake a cake. () P [ Q] T T VP will P bake they NP a N cake Verb bake needs

More information

Parsing. Based on presentations from Chris Manning s course on Statistical Parsing (Stanford)

Parsing. Based on presentations from Chris Manning s course on Statistical Parsing (Stanford) Parsing Based on presentations from Chris Manning s course on Statistical Parsing (Stanford) S N VP V NP D N John hit the ball Levels of analysis Level Morphology/Lexical POS (morpho-synactic), WSD Elements

More information

Dependency grammar. Recurrent neural networks. Transition-based neural parsing. Word representations. Informs Models

Dependency grammar. Recurrent neural networks. Transition-based neural parsing. Word representations. Informs Models Dependency grammar Morphology Word order Transition-based neural parsing Word representations Recurrent neural networks Informs Models Dependency grammar Morphology Word order Transition-based neural parsing

More information

Constituency. Doug Arnold

Constituency. Doug Arnold Constituency Doug Arnold doug@essex.ac.uk Spose we have a string... xyz..., how can we establish whether xyz is a constituent (i.e. syntactic unit); i.e. whether the representation of... xyz... should

More information

Ling 98a: The Meaning of Negation (Week 5)

Ling 98a: The Meaning of Negation (Week 5) Yimei Xiang yxiang@fas.harvard.edu 15 October 2013 1 Review Negation in propositional logic, oppositions, term logic of Aristotle Presuppositions Projection and accommodation Three-valued logic External/internal

More information

Unit 2: Tree Models. CS 562: Empirical Methods in Natural Language Processing. Lectures 19-23: Context-Free Grammars and Parsing

Unit 2: Tree Models. CS 562: Empirical Methods in Natural Language Processing. Lectures 19-23: Context-Free Grammars and Parsing CS 562: Empirical Methods in Natural Language Processing Unit 2: Tree Models Lectures 19-23: Context-Free Grammars and Parsing Oct-Nov 2009 Liang Huang (lhuang@isi.edu) Big Picture we have already covered...

More information

Models of Adjunction in Minimalist Grammars

Models of Adjunction in Minimalist Grammars Models of Adjunction in Minimalist Grammars Thomas Graf mail@thomasgraf.net http://thomasgraf.net Stony Brook University FG 2014 August 17, 2014 The Theory-Neutral CliffsNotes Insights Several properties

More information

HPSG II: the plot thickens

HPSG II: the plot thickens Syntactic Models 2/21/06 HPSG II: the plot thickens 1 Passive: a lexical rule that rearranges ARG-ST! (1) Passive Lexical Rule < 1, tv - lxm ARG - ST INDEX i < FPSP 1 a, > part - lxm SYN HEAD FORM pass

More information

Features. An argument DP must have been assigned Case by S-structure. A Specifier of IP must have been occupied by something by S-structure.

Features. An argument DP must have been assigned Case by S-structure. A Specifier of IP must have been occupied by something by S-structure. Features 26 October 2015 We left with a system that included the following elements. (1) Head Movement onstraint IfanX 0 movestoay 0,therecanbenoZ 0 thatisc-commandedbyy 0 and c-commandsx 0. (2) Like(s)

More information

Some binding facts. Binding in HPSG. The three basic principles. Binding theory of Chomsky

Some binding facts. Binding in HPSG. The three basic principles. Binding theory of Chomsky Some binding facts Binding in HPSG Binding theory has to do with constraining referentially dependent elements in terms of what is a possible antecedent, or binder. Introduction to HPSG June, 009 (1) John

More information

Recap: Lexicalized PCFGs (Fall 2007): Lecture 5 Parsing and Syntax III. Recap: Charniak s Model. Recap: Adding Head Words/Tags to Trees

Recap: Lexicalized PCFGs (Fall 2007): Lecture 5 Parsing and Syntax III. Recap: Charniak s Model. Recap: Adding Head Words/Tags to Trees Recap: Lexicalized PCFGs We now need to estimate rule probabilities such as P rob(s(questioned,vt) NP(lawyer,NN) VP(questioned,Vt) S(questioned,Vt)) 6.864 (Fall 2007): Lecture 5 Parsing and Syntax III

More information

Parsing with Context-Free Grammars

Parsing with Context-Free Grammars Parsing with Context-Free Grammars CS 585, Fall 2017 Introduction to Natural Language Processing http://people.cs.umass.edu/~brenocon/inlp2017 Brendan O Connor College of Information and Computer Sciences

More information

CS 188 Introduction to AI Fall 2005 Stuart Russell Final

CS 188 Introduction to AI Fall 2005 Stuart Russell Final NAME: SID#: Section: 1 CS 188 Introduction to AI all 2005 Stuart Russell inal You have 2 hours and 50 minutes. he exam is open-book, open-notes. 100 points total. Panic not. Mark your answers ON HE EXAM

More information

HPSG: Binding Theory

HPSG: Binding Theory HPSG: Binding Theory Doug Arnold doug@essexacuk Introduction Binding Theory is to do with the syntactic restrictions on the distribution of referentially dependent items and their antecedents: reflexives/reciprocals

More information

CS460/626 : Natural Language

CS460/626 : Natural Language CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 23, 24 Parsing Algorithms; Parsing in case of Ambiguity; Probabilistic Parsing) Pushpak Bhattacharyya CSE Dept., IIT Bombay 8 th,

More information

A Context-Free Grammar

A Context-Free Grammar Statistical Parsing A Context-Free Grammar S VP VP Vi VP Vt VP VP PP DT NN PP PP P Vi sleeps Vt saw NN man NN dog NN telescope DT the IN with IN in Ambiguity A sentence of reasonable length can easily

More information

Natural Language Processing : Probabilistic Context Free Grammars. Updated 5/09

Natural Language Processing : Probabilistic Context Free Grammars. Updated 5/09 Natural Language Processing : Probabilistic Context Free Grammars Updated 5/09 Motivation N-gram models and HMM Tagging only allowed us to process sentences linearly. However, even simple sentences require

More information

Recap: Tree geometry, selection, Θ-theory

Recap: Tree geometry, selection, Θ-theory Syntax II Seminar 2 Recap: Tree geometry, selection, Θ-theory Dr. James Grifiihs james.grifiihs@uni-konsianz.de Recap: tree geometry terminology M N O D E F H I J Branches 2 Recap: tree geometry terminology

More information

The Formal Architecture of. Lexical-Functional Grammar. Ronald M. Kaplan and Mary Dalrymple

The Formal Architecture of. Lexical-Functional Grammar. Ronald M. Kaplan and Mary Dalrymple The Formal Architecture of Lexical-Functional Grammar Ronald M. Kaplan and Mary Dalrymple Xerox Palo Alto Research Center 1. Kaplan and Dalrymple, ESSLLI 1995, Barcelona Architectural Issues Representation:

More information

Grundlagenmodul Semantik All Exercises

Grundlagenmodul Semantik All Exercises Grundlagenmodul Semantik All Exercises Sommersemester 2014 Exercise 1 Are the following statements correct? Justify your answers in a single short sentence. 1. 11 {x x is a square number} 2. 11 {x {y y

More information

An introduction to German Syntax. 1. Head directionality: A major source of linguistic divergence

An introduction to German Syntax. 1. Head directionality: A major source of linguistic divergence An introduction to German Syntax 19 January 2018 1. Head directionality: A major source of linguistic divergence In English, heads uniformly precede their complements: (1) a. [ kiss Mary] a. * [ Mary kiss]

More information

Class Notes: Tsujimura (2007), Ch. 5. Syntax (1), pp (3) a. [[akai hon]-no hyooshi] b. [akai [hon-no hyooshi]]

Class Notes: Tsujimura (2007), Ch. 5. Syntax (1), pp (3) a. [[akai hon]-no hyooshi] b. [akai [hon-no hyooshi]] Class otes: Tsujimura (2007), Ch. 5. yntax (1), pp. 206-220 p. 206 What is!ytx"?! n area in linguistics that deals with the REGULRLITY of how words are put together to create grammatical sentences What

More information

Handout 8: Computation & Hierarchical parsing II. Compute initial state set S 0 Compute initial state set S 0

Handout 8: Computation & Hierarchical parsing II. Compute initial state set S 0 Compute initial state set S 0 Massachusetts Institute of Technology 6.863J/9.611J, Natural Language Processing, Spring, 2001 Department of Electrical Engineering and Computer Science Department of Brain and Cognitive Sciences Handout

More information

Sharpening the empirical claims of generative syntax through formalization

Sharpening the empirical claims of generative syntax through formalization Sharpening the empirical claims of generative syntax through formalization Tim Hunter University of Minnesota, Twin Cities ESSLLI, August 2015 Part 1: Grammars and cognitive hypotheses What is a grammar?

More information

Time Zones - KET Grammar

Time Zones - KET Grammar Inventory of grammatical areas Verbs Regular and irregular forms Pages 104-105 (Unit 1 Modals can (ability; requests; permission) could (ability; polite requests) Page 3 (Getting Started) Pages 45-47,

More information

2013 ISSN: JATLaC Journal 8: t 1. t t Chomsky 1993 I Radford (2009) R I t t R I 2. t R t (1) (= R's (15), p. 86) He could have helped

2013 ISSN: JATLaC Journal 8: t 1. t t Chomsky 1993 I Radford (2009) R I t t R I 2. t R t (1) (= R's (15), p. 86) He could have helped t 1. tt Chomsky 1993 IRadford (2009) R Itt R I 2. t R t (1) (= R's (15), p. 86) He could have helped her, or [she have helped him]. 2 have I has had I I could ellipsis I gapping R (2) (= R (18), p.88)

More information

Stepanov 2007: The End of CED? Minimalism and Extraction Domains

Stepanov 2007: The End of CED? Minimalism and Extraction Domains Stepanov (2007) Stepanov 2007: The End of CED? Minimalism and Extraction Domains 1 Introduction In English (and other languages), overt wh-extraction out of subjects or adjuncts (as opposed to objects)

More information

CS626: NLP, Speech and the Web. Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 14: Parsing Algorithms 30 th August, 2012

CS626: NLP, Speech and the Web. Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 14: Parsing Algorithms 30 th August, 2012 CS626: NLP, Speech and the Web Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 14: Parsing Algorithms 30 th August, 2012 Parsing Problem Semantics Part of Speech Tagging NLP Trinity Morph Analysis

More information

Introduction to Semantics. Common Nouns and Adjectives in Predicate Position 1

Introduction to Semantics. Common Nouns and Adjectives in Predicate Position 1 Common Nouns and Adjectives in Predicate Position 1 (1) The Lexicon of Our System at Present a. Proper Names: [[ Barack ]] = Barack b. Intransitive Verbs: [[ smokes ]] = [ λx : x D e. IF x smokes THEN

More information

S NP VP 0.9 S VP 0.1 VP V NP 0.5 VP V 0.1 VP V PP 0.1 NP NP NP 0.1 NP NP PP 0.2 NP N 0.7 PP P NP 1.0 VP NP PP 1.0. N people 0.

S NP VP 0.9 S VP 0.1 VP V NP 0.5 VP V 0.1 VP V PP 0.1 NP NP NP 0.1 NP NP PP 0.2 NP N 0.7 PP P NP 1.0 VP  NP PP 1.0. N people 0. /6/7 CS 6/CS: Natural Language Processing Instructor: Prof. Lu Wang College of Computer and Information Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang The grammar: Binary, no epsilons,.9..5

More information

Semantics and Generative Grammar. Quantificational DPs, Part 3: Covert Movement vs. Type Shifting 1

Semantics and Generative Grammar. Quantificational DPs, Part 3: Covert Movement vs. Type Shifting 1 Quantificational DPs, Part 3: Covert Movement vs. Type Shifting 1 1. Introduction Thus far, we ve considered two competing analyses of sentences like those in (1). (1) Sentences Where a Quantificational

More information

Uniformity of Theta- Assignment Hypothesis. Transit i ve s. Unaccusatives. Unergatives. Double object constructions. Previously... Bill lied.

Uniformity of Theta- Assignment Hypothesis. Transit i ve s. Unaccusatives. Unergatives. Double object constructions. Previously... Bill lied. Preiously... Uniformity of Theta- Assignment Hypothesis, daughter of P = Agent, daughter of P = Theme PP, daughter of! = Goal P called P Chris P P books gae P to PP Chris Unaccusaties Transit i e s The

More information

CMPT-825 Natural Language Processing. Why are parsing algorithms important?

CMPT-825 Natural Language Processing. Why are parsing algorithms important? CMPT-825 Natural Language Processing Anoop Sarkar http://www.cs.sfu.ca/ anoop October 26, 2010 1/34 Why are parsing algorithms important? A linguistic theory is implemented in a formal system to generate

More information

Intellectual Property of Mariamalia Hidalgo

Intellectual Property of Mariamalia Hidalgo Grammar Review Present Tense of BE Present I am ( m) am not you are ( re) aren t Affirmative sentences Subject + verb be + object. he is ( s) isn t sentences Subject + (verb be + not) + object. she is

More information

Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing

Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing Natural Language Processing! CS 6120 Spring 2014! Northeastern University!! David Smith! with some slides from Jason Eisner & Andrew

More information

Hierarchical and Linear Constraints on Structure

Hierarchical and Linear Constraints on Structure Hierarchical and Linear Constraints on Structure Peter Sells Sandbjerg Ph.D. course June 14 16, 2006 Part II 1 OT Evaluations (1) CP XP C C IP Pro/Adv NP Adv I I Adv/Neg/NQ/Pro VP V Pro V NP NP PP V Prt

More information

Proseminar on Semantic Theory Fall 2013 Ling 720 The Proper Treatment of Quantification in Ordinary English, Part 1: The Fragment of English

Proseminar on Semantic Theory Fall 2013 Ling 720 The Proper Treatment of Quantification in Ordinary English, Part 1: The Fragment of English The Proper Treatment of Quantification in Ordinary English, Part 1: The Fragment of English We will now explore the analysis of English that Montague puts forth in his seminal paper, PTQ. As we ve already

More information

Categories and Transformations 321

Categories and Transformations 321 Categories and ransformations 321 4.10 Functional Categories and Formal Features What precedes substantially revises the framework developed in chapters 1 3. But we have not yet subjected functional categories

More information

CS 662 Sample Midterm

CS 662 Sample Midterm Name: 1 True/False, plus corrections CS 662 Sample Midterm 35 pts, 5 pts each Each of the following statements is either true or false. If it is true, mark it true. If it is false, correct the statement

More information

CS 6120/CS4120: Natural Language Processing

CS 6120/CS4120: Natural Language Processing CS 6120/CS4120: Natural Language Processing Instructor: Prof. Lu Wang College of Computer and Information Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang Assignment/report submission

More information

Introduction to Semantics. Pronouns and Variable Assignments. We ve seen that implicatures are crucially related to context.

Introduction to Semantics. Pronouns and Variable Assignments. We ve seen that implicatures are crucially related to context. Pronouns and Variable Assignments 1. Putting this Unit in Context (1) What We ve Done So Far This Unit Expanded our semantic theory so that it includes (the beginnings of) a theory of how the presuppositions

More information

Unterspezifikation in der Semantik Scope Semantics in Lexicalized Tree Adjoining Grammars

Unterspezifikation in der Semantik Scope Semantics in Lexicalized Tree Adjoining Grammars in der emantik cope emantics in Lexicalized Tree Adjoining Grammars Laura Heinrich-Heine-Universität Düsseldorf Wintersemester 2011/2012 LTAG: The Formalism (1) Tree Adjoining Grammars (TAG): Tree-rewriting

More information

Semantics and Generative Grammar. Pronouns and Variable Assignments 1. We ve seen that implicatures are crucially related to context.

Semantics and Generative Grammar. Pronouns and Variable Assignments 1. We ve seen that implicatures are crucially related to context. Pronouns and Variable Assignments 1 1. Putting this Unit in Context (1) What We ve Done So Far This Unit Expanded our semantic theory so that it includes (the beginnings of) a theory of how the presuppositions

More information

Spring 2017 Ling 620. An Introduction to the Semantics of Tense 1

Spring 2017 Ling 620. An Introduction to the Semantics of Tense 1 1. Introducing Evaluation Times An Introduction to the Semantics of Tense 1 (1) Obvious, Fundamental Fact about Sentences of English The truth of some sentences (of English) depends upon the time they

More information

Spring 2018 Ling 620 Introduction to Semantics of Questions: Questions as Sets of Propositions (Hamblin 1973, Karttunen 1977)

Spring 2018 Ling 620 Introduction to Semantics of Questions: Questions as Sets of Propositions (Hamblin 1973, Karttunen 1977) Introduction to Semantics of Questions: Questions as Sets of Propositions (Hamblin 1973, Karttunen 1977) 1. Question Meanings and Sets of Propositions (1) The Semantics of Declarative Sentence Dave smokes

More information

Computational Linguistics

Computational Linguistics Computational Linguistics Dependency-based Parsing Clayton Greenberg Stefan Thater FR 4.7 Allgemeine Linguistik (Computerlinguistik) Universität des Saarlandes Summer 2016 Acknowledgements These slides

More information

Computational Linguistics. Acknowledgements. Phrase-Structure Trees. Dependency-based Parsing

Computational Linguistics. Acknowledgements. Phrase-Structure Trees. Dependency-based Parsing Computational Linguistics Dependency-based Parsing Dietrich Klakow & Stefan Thater FR 4.7 Allgemeine Linguistik (Computerlinguistik) Universität des Saarlandes Summer 2013 Acknowledgements These slides

More information

Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing

Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing Natural Language Processing CS 4120/6120 Spring 2017 Northeastern University David Smith with some slides from Jason Eisner & Andrew

More information

Simpler Syntax. Ling : Sign-Based Construction Grammar Instructor: Ivan A. Sag URL:

Simpler Syntax. Ling : Sign-Based Construction Grammar Instructor: Ivan A. Sag URL: Simpler Syntax Ling 7800-065: Sign-Based Construction Grammar Instructor: Ivan A. Sag (sag@stanford.edu) URL: http://lingo.stanford.edu/sag/li11-sbcg 1/ 58 Constructs ] [mtr sign construct : dtrs nelist(sign)

More information

BARE PHRASE STRUCTURE, LCA AND LINEARIZATION

BARE PHRASE STRUCTURE, LCA AND LINEARIZATION Neuere Entwicklungen der Grammatiktheorie Universität Wien, SoSe 2008 Winfried Lechner wlechner@gs.uoa.gr Handout #2 BARE PHRASE STRUCTURE, LCA AND LINEARIZATION http://vivaldi.sfs.nphil.uni-tuebingen.de/%7ennsle01/synvie.htm

More information

Control and Tough- Movement

Control and Tough- Movement Control and Tough- Movement Carl Pollard February 2, 2012 Control (1/5) We saw that PRO is used for the unrealized subject of nonfinite verbals and predicatives where the subject plays a semantic role

More information

Structures mathématiques du langage

Structures mathématiques du langage tructures mathématiques du langage Alain Lecomte 16 février 2014 1 Heim and Kratzer s theory Montague s grammar was conceived and built during the sixties of the last century, without much attention paid

More information

DT2118 Speech and Speaker Recognition

DT2118 Speech and Speaker Recognition DT2118 Speech and Speaker Recognition Language Modelling Giampiero Salvi KTH/CSC/TMH giampi@kth.se VT 2015 1 / 56 Outline Introduction Formal Language Theory Stochastic Language Models (SLM) N-gram Language

More information

CS 224N HW:#3. (V N0 )δ N r p r + N 0. N r (r δ) + (V N 0)δ. N r r δ. + (V N 0)δ N = 1. 1 we must have the restriction: δ NN 0.

CS 224N HW:#3. (V N0 )δ N r p r + N 0. N r (r δ) + (V N 0)δ. N r r δ. + (V N 0)δ N = 1. 1 we must have the restriction: δ NN 0. CS 224 HW:#3 ARIA HAGHIGHI SUID :# 05041774 1. Smoothing Probability Models (a). Let r be the number of words with r counts and p r be the probability for a word with r counts in the Absolute discounting

More information

Computer Sciences Department

Computer Sciences Department Computer Sciences Department 1 Reference Book: INTRODUCTION TO THE THEORY OF COMPUTATION, SECOND EDITION, by: MICHAEL SIPSER Computer Sciences Department 3 ADVANCED TOPICS IN C O M P U T A B I L I T Y

More information

Chiastic Lambda-Calculi

Chiastic Lambda-Calculi Chiastic Lambda-Calculi wren ng thornton Cognitive Science & Computational Linguistics Indiana University, Bloomington NLCS, 28 June 2013 wren ng thornton (Indiana University) Chiastic Lambda-Calculi NLCS,

More information

Constituency Parsing

Constituency Parsing CS5740: Natural Language Processing Spring 2017 Constituency Parsing Instructor: Yoav Artzi Slides adapted from Dan Klein, Dan Jurafsky, Chris Manning, Michael Collins, Luke Zettlemoyer, Yejin Choi, and

More information

Model-Theory of Property Grammars with Features

Model-Theory of Property Grammars with Features Model-Theory of Property Grammars with Features Denys Duchier Thi-Bich-Hanh Dao firstname.lastname@univ-orleans.fr Yannick Parmentier Abstract In this paper, we present a model-theoretic description of

More information

Two-phase Implementation of Morphological Analysis

Two-phase Implementation of Morphological Analysis Two-phase Implementation of Morphological Analysis Arvi Hurskainen Institute for Asian and African Studies, Box 59 FIN-00014 University of Helsinki, Finland arvi.hurskainen@helsinki.fi Abstract SALAMA

More information

564 Lecture 25 Nov. 23, Continuing note on presuppositional vs. nonpresuppositional dets.

564 Lecture 25 Nov. 23, Continuing note on presuppositional vs. nonpresuppositional dets. 564 Lecture 25 Nov. 23, 1999 1 Continuing note on presuppositional vs. nonpresuppositional dets. Here's the argument about the nonpresupp vs. presupp analysis of "every" that I couldn't reconstruct last

More information

Control and Tough- Movement

Control and Tough- Movement Department of Linguistics Ohio State University February 2, 2012 Control (1/5) We saw that PRO is used for the unrealized subject of nonfinite verbals and predicatives where the subject plays a semantic

More information

Spring 2018 Ling 620 The Basics of Intensional Semantics, Part 1: The Motivation for Intensions and How to Formalize Them 1

Spring 2018 Ling 620 The Basics of Intensional Semantics, Part 1: The Motivation for Intensions and How to Formalize Them 1 The Basics of Intensional Semantics, Part 1: The Motivation for Intensions and How to Formalize Them 1 1. The Inadequacies of a Purely Extensional Semantics (1) Extensional Semantics a. The interpretation

More information

Ling 5801: Lecture Notes 7 From Programs to Context-Free Grammars

Ling 5801: Lecture Notes 7 From Programs to Context-Free Grammars Ling 5801: Lecture otes 7 From Programs to Context-Free rammars 1. The rules we used to define programs make up a context-free grammar A Context-Free rammar is a tuple C,X,S,R, where: C is a finite set

More information

POS-Tagging. Fabian M. Suchanek

POS-Tagging. Fabian M. Suchanek POS-Tagging Fabian M. Suchanek 100 Def: POS A Part-of-Speech (also: POS, POS-tag, word class, lexical class, lexical category) is a set of words with the same grammatical role. Alizée wrote a really great

More information

Dialogue Systems. Statistical NLU component. Representation. A Probabilistic Dialogue System. Task: map a sentence + context to a database query

Dialogue Systems. Statistical NLU component. Representation. A Probabilistic Dialogue System. Task: map a sentence + context to a database query Statistical NLU component Task: map a sentence + context to a database query Dialogue Systems User: Show me flights from NY to Boston, leaving tomorrow System: [returns a list of flights] Origin (City

More information

Context- Free Parsing with CKY. October 16, 2014

Context- Free Parsing with CKY. October 16, 2014 Context- Free Parsing with CKY October 16, 2014 Lecture Plan Parsing as Logical DeducBon Defining the CFG recognibon problem BoHom up vs. top down Quick review of Chomsky normal form The CKY algorithm

More information

One hint from secondary predication (from Baker 1997 (8) A secondary predicate cannot take the goal argument as subject of predication, wheth

One hint from secondary predication (from Baker 1997 (8) A secondary predicate cannot take the goal argument as subject of predication, wheth MIT, Fall 2003 1 The Double Object Construction (Larson 1988, Aoun & Li 1989) MIT, 24.951, Fr 14 Nov 2003 A familiar puzzle The Dative Alternation (1) a. I gave the candy to the children b. I gave the

More information

Natural Language Processing

Natural Language Processing SFU NatLangLab Natural Language Processing Anoop Sarkar anoopsarkar.github.io/nlp-class Simon Fraser University September 27, 2018 0 Natural Language Processing Anoop Sarkar anoopsarkar.github.io/nlp-class

More information

Probabilistic Context-free Grammars

Probabilistic Context-free Grammars Probabilistic Context-free Grammars Computational Linguistics Alexander Koller 24 November 2017 The CKY Recognizer S NP VP NP Det N VP V NP V ate NP John Det a N sandwich i = 1 2 3 4 k = 2 3 4 5 S NP John

More information

Logical Translations Jean Mark Gawron San Diego State University. 1 Introduction 2

Logical Translations Jean Mark Gawron San Diego State University. 1 Introduction 2 Logical Translations Jean Mark Gawron San Diego State University Contents 1 Introduction 2 2 Truth-Functional Connectives 2 2.1 And................................ 2 2.2 Or.................................

More information

10/17/04. Today s Main Points

10/17/04. Today s Main Points Part-of-speech Tagging & Hidden Markov Model Intro Lecture #10 Introduction to Natural Language Processing CMPSCI 585, Fall 2004 University of Massachusetts Amherst Andrew McCallum Today s Main Points

More information

Syntax and Semantics in Minimalist Grammars

Syntax and Semantics in Minimalist Grammars Syntax and Semantics in Minimalist Grammars Gregory M. Kobele ESSLLI 09 This chapter presents a grammar for a fragment of English A-movement. 1 The constructions accounted for include raising, passivization,

More information

Handout 3: PTQ Revisited (Muskens 1995, Ch. 4)

Handout 3: PTQ Revisited (Muskens 1995, Ch. 4) Handout 3: PTQ Revisited (Muskens 1995, Ch. 4) Semantics C (Spring 2010) Montague s PTQ article (Montague 1973), the paper in which he gave his Proper Treatment of Quantification, functions as the paradigm

More information

Grammar and Feature Unification

Grammar and Feature Unification Grammar and Feature Unification Problems with CF Phrase Structure Grammars Difficult to capture dependencies between constituents the boy runs the boys run * the boy run * the boys runs Problems with CF

More information

Tree Adjoining Grammars

Tree Adjoining Grammars Tree Adjoining Grammars Feature Structure Based TAG Laura Kallmeyer & Benjamin Burkhardt HHU Düsseldorf WS 2017/2018 1 / 20 Outline 1 Why feature structures? 2 Basics of feature structure logic 3 Feature

More information

Two Reconstruction Puzzles Yael Sharvit University of Connecticut

Two Reconstruction Puzzles Yael Sharvit University of Connecticut Workshop on Direct Compositionality June 19-21, 2003 Brown University Two Reconstruction Puzzles Yael Sharvit University of Connecticut yael.sharvit@uconn.edu Some constructions exhibit what is known as

More information

Parasitic Scope (Barker 2007) Semantics Seminar 11/10/08

Parasitic Scope (Barker 2007) Semantics Seminar 11/10/08 Parasitic Scope (Barker 2007) Semantics Seminar 11/10/08 1. Overview Attempts to provide a compositional, fully semantic account of same. Elements other than NPs in particular, adjectives can be scope-taking

More information

Review. Earley Algorithm Chapter Left Recursion. Left-Recursion. Rule Ordering. Rule Ordering

Review. Earley Algorithm Chapter Left Recursion. Left-Recursion. Rule Ordering. Rule Ordering Review Earley Algorithm Chapter 13.4 Lecture #9 October 2009 Top-Down vs. Bottom-Up Parsers Both generate too many useless trees Combine the two to avoid over-generation: Top-Down Parsing with Bottom-Up

More information