X-bar theory. X-bar :

Size: px
Start display at page:

Download "X-bar theory. X-bar :"

Transcription

1 is one of the greatest contributions of generative school in the filed of knowledge system. Besides linguistics, computer science is greatly indebted to Chomsky to have propounded the theory of x-bar. X-bar is a theory because it is testable and it captures the generalization by bringing the desired results which are systematic in nature. If you recall, we have had instances, for example, in case of c-command, domination, precedence and even in government, where we talked about the intervening nodes between the two nodes. If one tries to sum up the essence of x-bar in one sentence, s/he would like to say that.it facilitates the generative grammar to explain/capture and describe this intervening node by providing an intermediate node for the configurational relationship in a phrase. This explanation may not look convincing at the outset of the discussion, but we will gradually realize that what we have said about the x-bar is very logical and explanatory in nature!

2 Carnie (2012;167) gives a bibliographic note about x-bar theory. He says that the first presentation of appeared in Chomsky (1970). However, Jackendoff s (1977) seminal book X-bar Syntax is the source of many of the ideas surrounding. If one wants to know the best account of x-bar theory one should consult Radford s (1988) Transformational Grammar: A First Course. This textbook presents one of the most comprehensive arguments for. So, as we mentioned earlier, the need of an intermediate node (pronounced bar ) is the NEED of the hour in order to explain the items that are conjoined in the internal structure of sentences and they must be explained with care. In other words, every node does not have one and same relationship in a tree-diagram, and if we want to distinguish these differences, we must avoid the flat structure of a sentence, where all the attachments in a tree diagram seem to bear similar relationship. The hierarchy in a tree-diagram can be explained better if we have a binary-branching for the bifurcation of the nodes (branches) in a tree-diagram. The helps in doing so and it is a means of the original mechanism for indicating intermediate categories in a tree-diagram.

3 . X-bar : 1. NP (D) N' 2. N' (AdjP) N' or N' (PP) 3. N' N (PP) 4. VP V' 5. V' V' (PP) or V' (AdvP) 6. V' V (NP) 7. AdvP Adv' 8. Adv' (AdvP) Adv' 9. Adv' Adv (PP) 10. AdjP Adj' 11. Adj' (AdvP) Adj' 12. Adj' Adj (PP) 13. PP P' 14. P' P' (PP) or (AdvP) P' 15. P' P (NP)

4 It is important to notice that in all the rules in the system that are formulated for the category is the same. In every rule, there is ONE element that is a MUST and it can t be optional. For example, in the NP rule (the first rule in our list), the element that isn t optional is N' (i.e. en-bar or N-bar). Similarly, the only obligatory element in N' is either another N' or N. This is a very general yet important notion in x-bar structure. We call it headedness, and all phrases appear to have heads. A head is the most prominent element in a phrasal category and gives its grammatical category to the whole phrase. Note that we don t have any rule as given below: * NP V AdjP This rule not only seems meaningless, it is unattested in the system we are trying to developed here. We have learned that the requirement of endocentricity demands that every phrase must have head. Thus, the only obligatory element in a phrase is the head.

5 We can condense the rules that we have proposed into a simple set. To do this we are going to make use of variables (like variables in algebra) to stand for particular parts of speech. Let X be a variable that can stand for any category N, V, Adj, Adv, P. XP is an umbrella-term that covers all the phrases such as NP, VP, AP, and PP. Similarly X' stands for an intermediate node that hasn t yet terminated into lexical items and they are thus called bar-level N', V', Adj', Adv', and P', and Finally X represents the terminal nodes which are lexical items such as N, V, Adj, Adv, and P, and they are the heads of their respective phrases. So, with this three-tiered mechanism inbuilt in, the is an attempt to capture these similarities among rules (i.e. for each kind of phrase, the same kind of rules should appear).

6 Please note that the non-head material in all the rules is not only phrasal but also optional (with the exception of the determiner in the NP rule for time being) So, we never find a rule like this: * V' Adv V Therefore, anything in an that isn t a head must be a phrase and optional (except NP rule for time being). Also notice that for each major category there are three rules; One that introduces the NP, VP, AdvP, AdjP, and PP, One that takes a bar level and repeats it (e.g., N' N' (PP)), and The one that takes a bar level and spells out the head (e.g., N' N (PP)). So, these rules should be viewed as; 1. XP (YP) X' 2. X' X' (ZP) or X' (ZP) X' 3. X' X (WP)

7 Let s now examine as to how should we explained different layers of the strucures of a phrase under the new schema; 1.a. a) N' N (PP) b) V' V (NP) c) Adj' Adj (PP) d) Adv' Adv (PP) e) P' P (NP) 2.a. a) N' (AdjP) N' or N' (PP) b) V' V' (PP) or V' (AdvP) c) Adj' (AdvP) Adj' d) Adv' (AdvP) Adv' e) P' P' (PP) or (AdvP) P' 1.b. By using the variable notation we can generalize across these rules with the single general statement: 2.b. Use single general statement: 1.c. X' X (WP)(to be revised) 2.c. X' (ZP) X' or X' (ZP) (to be revised)

8 Finally, let s consider the rule that introduces the topmost layer of the structure our phrase under the new schema; 3.a. a) NP (D) N' DP D' (NP) b) VP V' (NP) c) AdjP Adj' (AdjP) d) AdvP Adv' (AdvP) e) PP P' (NP) 3.b. By using the variable notation we can generalize across these rules with the single general statement: 3.c. XP (YP) X' (to be revised)

9 Back to the trees: X-bar Theory Consider our current NP rule: NP: (D) (AdjP+) N (PP+) This yields a flat structure where all of the components of DP c-command each other. NP D AdjP N PP PP this Adj book P NP P NP big of N with D AdjP N poems the Adj cover blue

10 X-bar Theory: NP I bought this big book of poems with the blue cover. You bought this small one. NP D AdjP N PP PP this Adj book P NP P NP big of N with D AdjP N poems the Adj cover blue

11 X-bar Theory: NP We can substitute one for book of poems with the blue cover, which should mean book of poems with the blue cover is a constituent, but it isn t in our structure. NP D AdjP N PP PP this Adj book P NP P NP big of N with D AdjP N poems the Adj cover blue

12 X-bar Theory: NP I bought this small one with the red cover. We can also substitute one in for book of poems alone, which should thus also be a constituent. NP D AdjP N PP PP this Adj book P NP P NP big of N with D AdjP N poems the Adj cover blue

13 X-bar Theory: NP This suggests a more deeply embedded structure: NP?? D AdjP N PP PP this Adj book P NP P NP big of N with D AdjP N poems the Adj cover blue

14 X-bar Theory: NP We ll call these intermediate nodes of NP N (N-bar). Notice that you can also say I bought this one. NP N N D AdjP N PP PP this Adj book P NP P NP big of N with D AdjP N poems the Adj cover blue

15 X-bar Theory: NP So, our final NP looks like this: NP N N N D AdjP N PP PP this Adj book P NP P NP big of N with D AdjP N poems the Adj cover blue

16 X-bar Theory: NP We need to break up our NP rule; instead of NP: (D) (AdjP+) N (PP+) We have: NP: (D) N N : AdjP N N : N PP N : N (PP) Notice that these yield the same results on the surface (note the recursion and the optionality) but produce different structures (in terms of constituency). Also notice that under the new schema, all nodes of NP have ONLY two daughters. This seems to be a big deal as it changes the tree-relationship into a binary branching.

17 X-bar Theory: VP The same kind of thing holds true for VP as well as NP. Instead of using one (which stands for N ) we can try doing replacements using do so, and we ll get a very similar result. Our old rule generated a flat structure for VP as well It had all PPs, NPs, CPs, etc. in a VP and they all c-commanded each other). VP: (AdvP+) V ({NP/CP}) (PP+) (AdvP+)

18 X-bar Theory: VP VP: (AdvP+) V ({NP/CP}) (PP+) (AdvP+) The chef ate the beans with a fork The chef quickly left after Mary did so. The chef left quickly after Mary did so. The chef ate the pizza with joy and Mary did so with quiet reverse. The chef ate the pizza with joy immediately and Mary did so later.

19 X-bar Theory: VP Again, it looks like we need to break our rule into parts using V (for which do so can substitute). VP: (AdvP+) V ({NP/CP}) (PP+) (AdvP+) To: VP: V V : AdvP V V : V PP V : V AdvP V : V ({NP/CP}) Again, this produces all the same on the surface, but yields a different structure. This again bifurcates the tree into a binary branching.

20 X-bar Theory: AdjP We should now be growing suspicious of our other rules, now that we have had to split up NP and VP and introduce N and V nodes. The governor was [ AdjP very concerned about housing costs ]. The director was [ AdjP unusually pleased with his actors and confident of success ]. This gives us evidence of AdjP: (AdvP) Adj Adj : Adj (PP)

21 X-bar Theory: PP The Frisbee landed on the roof. It landed right on the edge. John knocked it right off the roof and into the trashcan. So, this gives us (assuming right is an AdjP): PP: (AdjP) P P : P (PP) P : P NP

22 The main idea behind is to explain the similarity between the rules for each category. It is an attempt to generalize over the rules we have. PP: (AdjP) P P : P (PP) P : P DP AdjP: (AdvP) Adj Adj : Adj (PP) NP: (D) N N : AdjP N N : N PP N : N (PP) VP: V V : AdvP V V : V PP V : V AdvP V : V ({NP/CP})

23 The X in is a variable over categories. When we talk of XP, we mean to be describing any kind of phrase (VP, NP, AdjP, AdvP, PP, TP, CP, ). PP: (AdjP) P P : P (PP) P : P DP AdjP: (AdvP) Adj Adj : Adj (PP) NP: (D) N N : AdjP N N : N PP N : N (PP) VP: V V : AdvP V V : V PP V : V AdvP V : V ({NP/CP})

24 All the rules all have the following form: XP: YP X X : (ZP) X / X : X (ZP) X : X (WP) PP: (AdjP) P P : P (PP) P : P DP AdjP: (AdvP) Adj Adj : Adj (PP) NP: (D) N N : AdjP N N : N PP N : N (PP) VP: V V : AdvP V V : V PP V : V AdvP V : V ({NP/CP})

25 elevates this to a principle of phrase structure; it hypothesizes that all phrases in a syntactic tree conform to this template. XP : (YP) X A phrase (XP) consists of optionally another phrase and a barlevel projection (X ). X : ZP X or X : X ZP A bar-level projection (X ) can consist of another X and another phrase (recursive). X : X (WP) A bar-level projection (X ) consists of a head of the same category (X) and optionally another phrase.

26 Structurally, this looks like this (of course, there can be any number of X nodes, here we see three). YP XP X Different parts of this structure are given different names (and they act different from one another, as we ll see). ZP X X X WP ZP

27 The phrase which is immediately dominated by XP (designated YP here) is the specifier. YP XP X ZP X X ZP X WP

28 The phrase which is immediately dominated by XP (designated YP here) is the specifier. YP XP X A phrase dominated by X and the sister of X is an adjunct. ZP X X ZP X WP

29 The phrase which is immediately dominated by XP (designated YP here) is the specifier. YP XP X A phrase dominated by X and the sister of X is an adjunct. The phrase which is sister to X is the complement. ZP X X X WP ZP

Another look at PSRs: Intermediate Structure. Starting X-bar theory

Another look at PSRs: Intermediate Structure. Starting X-bar theory Another look at PSRs: Intermediate Structure Starting X-bar theory Andrew Carnie, 2006 Substitution Andrew Carnie, 2006 Substitution If a group of words can be replaced by a single word, they are a constituent.

More information

Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically

Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically Ch. 2: Phrase Structure Syntactic Structure (basic concepts) A tree diagram marks constituents hierarchically NP S AUX VP Ali will V NP help D N the man A node is any point in the tree diagram and it can

More information

CS 712: Topics in NLP Linguistic Phrases and Statistical Phrases

CS 712: Topics in NLP Linguistic Phrases and Statistical Phrases CS 712: Topics in NLP Linguistic Phrases and Statistical Phrases Pushpak Bhattacharyya, CSE Department, IIT Bombay 18 March, 2013 (main text: Syntax by Adrew Carnie, Blackwell Publication, 2002) Domination

More information

Andrew Carnie, Structural Relations. The mathematical properties of phrase structure trees

Andrew Carnie, Structural Relations. The mathematical properties of phrase structure trees Structural Relations The mathematical properties of phrase structure trees Important! Important! Even if you have trouble with the formal definitions, try to understand the INTUITIVE idea behind them.

More information

Artificial Intelligence

Artificial Intelligence CS344: Introduction to Artificial Intelligence Pushpak Bhattacharyya CSE Dept., IIT Bombay Lecture 20-21 Natural Language Parsing Parsing of Sentences Are sentences flat linear structures? Why tree? Is

More information

Models of Adjunction in Minimalist Grammars

Models of Adjunction in Minimalist Grammars Models of Adjunction in Minimalist Grammars Thomas Graf mail@thomasgraf.net http://thomasgraf.net Stony Brook University FG 2014 August 17, 2014 The Theory-Neutral CliffsNotes Insights Several properties

More information

(7) a. [ PP to John], Mary gave the book t [PP]. b. [ VP fix the car], I wonder whether she will t [VP].

(7) a. [ PP to John], Mary gave the book t [PP]. b. [ VP fix the car], I wonder whether she will t [VP]. CAS LX 522 Syntax I Fall 2000 September 18, 2000 Paul Hagstrom Week 2: Movement Movement Last time, we talked about subcategorization. (1) a. I can solve this problem. b. This problem, I can solve. (2)

More information

Sharpening the empirical claims of generative syntax through formalization

Sharpening the empirical claims of generative syntax through formalization Sharpening the empirical claims of generative syntax through formalization Tim Hunter University of Minnesota, Twin Cities ESSLLI, August 2015 Part 1: Grammars and cognitive hypotheses What is a grammar?

More information

Parsing with Context-Free Grammars

Parsing with Context-Free Grammars Parsing with Context-Free Grammars Berlin Chen 2005 References: 1. Natural Language Understanding, chapter 3 (3.1~3.4, 3.6) 2. Speech and Language Processing, chapters 9, 10 NLP-Berlin Chen 1 Grammars

More information

Introduction to Semantics. The Formalization of Meaning 1

Introduction to Semantics. The Formalization of Meaning 1 The Formalization of Meaning 1 1. Obtaining a System That Derives Truth Conditions (1) The Goal of Our Enterprise To develop a system that, for every sentence S of English, derives the truth-conditions

More information

2 A not-quite-argument for X-bar structure in noun phrases

2 A not-quite-argument for X-bar structure in noun phrases CAS LX 321 / GRS LX 621 Syntax: Introduction to Sentential Structure ovember 16, 2017 1 and pronouns (1) he linguists yodel. (2) We linguists yodel. (3) hey looked at us linguists. (4) hey looked at linguists.

More information

A* Search. 1 Dijkstra Shortest Path

A* Search. 1 Dijkstra Shortest Path A* Search Consider the eight puzzle. There are eight tiles numbered 1 through 8 on a 3 by three grid with nine locations so that one location is left empty. We can move by sliding a tile adjacent to the

More information

Ling 240 Lecture #15. Syntax 4

Ling 240 Lecture #15. Syntax 4 Ling 240 Lecture #15 Syntax 4 agenda for today Give me homework 3! Language presentation! return Quiz 2 brief review of friday More on transformation Homework 4 A set of Phrase Structure Rules S -> (aux)

More information

Sharpening the empirical claims of generative syntax through formalization

Sharpening the empirical claims of generative syntax through formalization Sharpening the empirical claims of generative syntax through formalization Tim Hunter University of Minnesota, Twin Cities NASSLLI, June 2014 Part 1: Grammars and cognitive hypotheses What is a grammar?

More information

CAS LX 522 Syntax I Fall 2000 October 10, 2000 Week 5: Case Theory and θ Theory. θ-theory continued

CAS LX 522 Syntax I Fall 2000 October 10, 2000 Week 5: Case Theory and θ Theory. θ-theory continued CAS LX 522 Syntax I Fall 2000 October 0, 2000 Paul Hagstrom Week 5: Case Theory and θ Theory θ-theory continued From last time: verbs have θ-roles (e.g., Agent, Theme, ) to assign, specified in the lexicon

More information

Semantics and Generative Grammar. A Little Bit on Adverbs and Events

Semantics and Generative Grammar. A Little Bit on Adverbs and Events A Little Bit on Adverbs and Events 1. From Adjectives to Adverbs to Events We ve just developed a theory of the semantics of adjectives, under which they denote either functions of type (intersective

More information

Recap: Tree geometry, selection, Θ-theory

Recap: Tree geometry, selection, Θ-theory Syntax II Seminar 2 Recap: Tree geometry, selection, Θ-theory Dr. James Grifiihs james.grifiihs@uni-konsianz.de Recap: tree geometry terminology M N O D E F H I J Branches 2 Recap: tree geometry terminology

More information

S NP VP 0.9 S VP 0.1 VP V NP 0.5 VP V 0.1 VP V PP 0.1 NP NP NP 0.1 NP NP PP 0.2 NP N 0.7 PP P NP 1.0 VP NP PP 1.0. N people 0.

S NP VP 0.9 S VP 0.1 VP V NP 0.5 VP V 0.1 VP V PP 0.1 NP NP NP 0.1 NP NP PP 0.2 NP N 0.7 PP P NP 1.0 VP  NP PP 1.0. N people 0. /6/7 CS 6/CS: Natural Language Processing Instructor: Prof. Lu Wang College of Computer and Information Science Northeastern University Webpage: www.ccs.neu.edu/home/luwang The grammar: Binary, no epsilons,.9..5

More information

Parsing. Based on presentations from Chris Manning s course on Statistical Parsing (Stanford)

Parsing. Based on presentations from Chris Manning s course on Statistical Parsing (Stanford) Parsing Based on presentations from Chris Manning s course on Statistical Parsing (Stanford) S N VP V NP D N John hit the ball Levels of analysis Level Morphology/Lexical POS (morpho-synactic), WSD Elements

More information

Lecture 7. Logic. Section1: Statement Logic.

Lecture 7. Logic. Section1: Statement Logic. Ling 726: Mathematical Linguistics, Logic, Section : Statement Logic V. Borschev and B. Partee, October 5, 26 p. Lecture 7. Logic. Section: Statement Logic.. Statement Logic..... Goals..... Syntax of Statement

More information

1. Background. Task: Determine whether a given string of words is a grammatical (well-formed) sentence of language L i or not.

1. Background. Task: Determine whether a given string of words is a grammatical (well-formed) sentence of language L i or not. Constraints in Syntax [1] Phrase Structure and Derivations Düsseldorf LSA/DGfS Summerschool 2002 Gereon Müller (IDS Mannheim) gereon.mueller@ids-mannheim.de 1. Background Task: Determine whether a given

More information

Uniformity of Theta- Assignment Hypothesis. Transit i ve s. Unaccusatives. Unergatives. Double object constructions. Previously... Bill lied.

Uniformity of Theta- Assignment Hypothesis. Transit i ve s. Unaccusatives. Unergatives. Double object constructions. Previously... Bill lied. Preiously... Uniformity of Theta- Assignment Hypothesis, daughter of P = Agent, daughter of P = Theme PP, daughter of! = Goal P called P Chris P P books gae P to PP Chris Unaccusaties Transit i e s The

More information

CAS LX 522 Syntax I November 4, 2002 Week 9: Wh-movement, supplement

CAS LX 522 Syntax I November 4, 2002 Week 9: Wh-movement, supplement CAS LX 522 Syntax I November 4, 2002 Fall 2002 Week 9: Wh-movement, supplement () Italian Tuo fratello ( your brother ), [ CP a cui i [ TP mi domando [ CP che storie i [ TP abbiano raccontato t i t j...

More information

A Context-Free Grammar

A Context-Free Grammar Statistical Parsing A Context-Free Grammar S VP VP Vi VP Vt VP VP PP DT NN PP PP P Vi sleeps Vt saw NN man NN dog NN telescope DT the IN with IN in Ambiguity A sentence of reasonable length can easily

More information

Some binding facts. Binding in HPSG. The three basic principles. Binding theory of Chomsky

Some binding facts. Binding in HPSG. The three basic principles. Binding theory of Chomsky Some binding facts Binding in HPSG Binding theory has to do with constraining referentially dependent elements in terms of what is a possible antecedent, or binder. Introduction to HPSG June, 009 (1) John

More information

CS460/626 : Natural Language

CS460/626 : Natural Language CS460/626 : Natural Language Processing/Speech, NLP and the Web (Lecture 23, 24 Parsing Algorithms; Parsing in case of Ambiguity; Probabilistic Parsing) Pushpak Bhattacharyya CSE Dept., IIT Bombay 8 th,

More information

Natural Language Processing : Probabilistic Context Free Grammars. Updated 5/09

Natural Language Processing : Probabilistic Context Free Grammars. Updated 5/09 Natural Language Processing : Probabilistic Context Free Grammars Updated 5/09 Motivation N-gram models and HMM Tagging only allowed us to process sentences linearly. However, even simple sentences require

More information

Parsing with CFGs L445 / L545 / B659. Dept. of Linguistics, Indiana University Spring Parsing with CFGs. Direction of processing

Parsing with CFGs L445 / L545 / B659. Dept. of Linguistics, Indiana University Spring Parsing with CFGs. Direction of processing L445 / L545 / B659 Dept. of Linguistics, Indiana University Spring 2016 1 / 46 : Overview Input: a string Output: a (single) parse tree A useful step in the process of obtaining meaning We can view the

More information

Parsing with CFGs. Direction of processing. Top-down. Bottom-up. Left-corner parsing. Chart parsing CYK. Earley 1 / 46.

Parsing with CFGs. Direction of processing. Top-down. Bottom-up. Left-corner parsing. Chart parsing CYK. Earley 1 / 46. : Overview L545 Dept. of Linguistics, Indiana University Spring 2013 Input: a string Output: a (single) parse tree A useful step in the process of obtaining meaning We can view the problem as searching

More information

Semantics 2 Part 1: Relative Clauses and Variables

Semantics 2 Part 1: Relative Clauses and Variables Semantics 2 Part 1: Relative Clauses and Variables Sam Alxatib EVELIN 2012 January 17, 2012 Reviewing Adjectives Adjectives are treated as predicates of individuals, i.e. as functions from individuals

More information

HPSG: Binding Theory

HPSG: Binding Theory HPSG: Binding Theory Doug Arnold doug@essexacuk Introduction Binding Theory is to do with the syntactic restrictions on the distribution of referentially dependent items and their antecedents: reflexives/reciprocals

More information

Computationele grammatica

Computationele grammatica Computationele grammatica Docent: Paola Monachesi Contents First Last Prev Next Contents 1 Unbounded dependency constructions (UDCs)............... 3 2 Some data...............................................

More information

LECTURER: BURCU CAN Spring

LECTURER: BURCU CAN Spring LECTURER: BURCU CAN 2017-2018 Spring Regular Language Hidden Markov Model (HMM) Context Free Language Context Sensitive Language Probabilistic Context Free Grammar (PCFG) Unrestricted Language PCFGs can

More information

CAS LX 523 Syntax II Spring 2001 March 13, (1) A qp. Kayne, Richard (1995). The antisymmetry of syntax. Cambridge, MA: MIT Press.

CAS LX 523 Syntax II Spring 2001 March 13, (1) A qp. Kayne, Richard (1995). The antisymmetry of syntax. Cambridge, MA: MIT Press. CAS LX 52 Syntax II Spring 200 March, 200 Paul Hagstrom Week 7: Antisymmetry Kayne, Richard (995). The antisymmetry of syntax. Cambridge, MA: MIT Press. Koopman, Hilda (2000). The spec-head configuration.

More information

Recap: Lexicalized PCFGs (Fall 2007): Lecture 5 Parsing and Syntax III. Recap: Charniak s Model. Recap: Adding Head Words/Tags to Trees

Recap: Lexicalized PCFGs (Fall 2007): Lecture 5 Parsing and Syntax III. Recap: Charniak s Model. Recap: Adding Head Words/Tags to Trees Recap: Lexicalized PCFGs We now need to estimate rule probabilities such as P rob(s(questioned,vt) NP(lawyer,NN) VP(questioned,Vt) S(questioned,Vt)) 6.864 (Fall 2007): Lecture 5 Parsing and Syntax III

More information

Natural Language Processing CS Lecture 06. Razvan C. Bunescu School of Electrical Engineering and Computer Science

Natural Language Processing CS Lecture 06. Razvan C. Bunescu School of Electrical Engineering and Computer Science Natural Language Processing CS 6840 Lecture 06 Razvan C. Bunescu School of Electrical Engineering and Computer Science bunescu@ohio.edu Statistical Parsing Define a probabilistic model of syntax P(T S):

More information

Control and Tough- Movement

Control and Tough- Movement Department of Linguistics Ohio State University February 2, 2012 Control (1/5) We saw that PRO is used for the unrealized subject of nonfinite verbals and predicatives where the subject plays a semantic

More information

cis32-ai lecture # 18 mon-3-apr-2006

cis32-ai lecture # 18 mon-3-apr-2006 cis32-ai lecture # 18 mon-3-apr-2006 today s topics: propositional logic cis32-spring2006-sklar-lec18 1 Introduction Weak (search-based) problem-solving does not scale to real problems. To succeed, problem

More information

Control and Tough- Movement

Control and Tough- Movement Control and Tough- Movement Carl Pollard February 2, 2012 Control (1/5) We saw that PRO is used for the unrealized subject of nonfinite verbals and predicatives where the subject plays a semantic role

More information

Unit 2: Tree Models. CS 562: Empirical Methods in Natural Language Processing. Lectures 19-23: Context-Free Grammars and Parsing

Unit 2: Tree Models. CS 562: Empirical Methods in Natural Language Processing. Lectures 19-23: Context-Free Grammars and Parsing CS 562: Empirical Methods in Natural Language Processing Unit 2: Tree Models Lectures 19-23: Context-Free Grammars and Parsing Oct-Nov 2009 Liang Huang (lhuang@isi.edu) Big Picture we have already covered...

More information

Proseminar on Semantic Theory Fall 2010 Ling 720. Remko Scha (1981/1984): Distributive, Collective and Cumulative Quantification

Proseminar on Semantic Theory Fall 2010 Ling 720. Remko Scha (1981/1984): Distributive, Collective and Cumulative Quantification 1. Introduction Remko Scha (1981/1984): Distributive, Collective and Cumulative Quantification (1) The Importance of Scha (1981/1984) The first modern work on plurals (Landman 2000) There are many ideas

More information

Top Down and Bottom Up Composition. 1 Notes on Notation and Terminology. 2 Top-down and Bottom-Up Composition. Two senses of functional application

Top Down and Bottom Up Composition. 1 Notes on Notation and Terminology. 2 Top-down and Bottom-Up Composition. Two senses of functional application Elizabeth Coppock Compositional Semantics coppock@phil.hhu.de Heinrich Heine University Wed. ovember 30th, 2011 Winter Semester 2011/12 Time: 14:30 16:00 Room: 25.22-U1.72 Top Down and Bottom Up Composition

More information

CHAPTER THREE: RELATIONS AND FUNCTIONS

CHAPTER THREE: RELATIONS AND FUNCTIONS CHAPTER THREE: RELATIONS AND FUNCTIONS 1 Relations Intuitively, a relation is the sort of thing that either does or does not hold between certain things, e.g. the love relation holds between Kim and Sandy

More information

CAS LX 522 Syntax I. We give trees to ditransitives. We give trees to ditransitives. We give trees to ditransitives. Problems continue UTAH (4.3-4.

CAS LX 522 Syntax I. We give trees to ditransitives. We give trees to ditransitives. We give trees to ditransitives. Problems continue UTAH (4.3-4. 7 CAS LX 522 Syntax I UTAH (4.3-4.4) You may recall our discussion of θ-theory, where we triumphantly classified erbs as coming in (at least) three types: Intransitie (1 θ-role) Transitie (2 θ-roles) Ditransitie

More information

Penn Treebank Parsing. Advanced Topics in Language Processing Stephen Clark

Penn Treebank Parsing. Advanced Topics in Language Processing Stephen Clark Penn Treebank Parsing Advanced Topics in Language Processing Stephen Clark 1 The Penn Treebank 40,000 sentences of WSJ newspaper text annotated with phrasestructure trees The trees contain some predicate-argument

More information

Semantics and Generative Grammar. Pronouns and Variable Assignments 1. We ve seen that implicatures are crucially related to context.

Semantics and Generative Grammar. Pronouns and Variable Assignments 1. We ve seen that implicatures are crucially related to context. Pronouns and Variable Assignments 1 1. Putting this Unit in Context (1) What We ve Done So Far This Unit Expanded our semantic theory so that it includes (the beginnings of) a theory of how the presuppositions

More information

Decoding and Inference with Syntactic Translation Models

Decoding and Inference with Syntactic Translation Models Decoding and Inference with Syntactic Translation Models March 5, 2013 CFGs S NP VP VP NP V V NP NP CFGs S NP VP S VP NP V V NP NP CFGs S NP VP S VP NP V NP VP V NP NP CFGs S NP VP S VP NP V NP VP V NP

More information

564 Lecture 25 Nov. 23, Continuing note on presuppositional vs. nonpresuppositional dets.

564 Lecture 25 Nov. 23, Continuing note on presuppositional vs. nonpresuppositional dets. 564 Lecture 25 Nov. 23, 1999 1 Continuing note on presuppositional vs. nonpresuppositional dets. Here's the argument about the nonpresupp vs. presupp analysis of "every" that I couldn't reconstruct last

More information

Introduction to Semantics. Pronouns and Variable Assignments. We ve seen that implicatures are crucially related to context.

Introduction to Semantics. Pronouns and Variable Assignments. We ve seen that implicatures are crucially related to context. Pronouns and Variable Assignments 1. Putting this Unit in Context (1) What We ve Done So Far This Unit Expanded our semantic theory so that it includes (the beginnings of) a theory of how the presuppositions

More information

Bringing machine learning & compositional semantics together: central concepts

Bringing machine learning & compositional semantics together: central concepts Bringing machine learning & compositional semantics together: central concepts https://githubcom/cgpotts/annualreview-complearning Chris Potts Stanford Linguistics CS 244U: Natural language understanding

More information

Introduction to Semantic Parsing with CCG

Introduction to Semantic Parsing with CCG Introduction to Semantic Parsing with CCG Kilian Evang Heinrich-Heine-Universität Düsseldorf 2018-04-24 Table of contents 1 Introduction to CCG Categorial Grammar (CG) Combinatory Categorial Grammar (CCG)

More information

1 Boolean Algebra Simplification

1 Boolean Algebra Simplification cs281: Computer Organization Lab3 Prelab Our objective in this prelab is to lay the groundwork for simplifying boolean expressions in order to minimize the complexity of the resultant digital logic circuit.

More information

CMPT-825 Natural Language Processing. Why are parsing algorithms important?

CMPT-825 Natural Language Processing. Why are parsing algorithms important? CMPT-825 Natural Language Processing Anoop Sarkar http://www.cs.sfu.ca/ anoop October 26, 2010 1/34 Why are parsing algorithms important? A linguistic theory is implemented in a formal system to generate

More information

Model-Theory of Property Grammars with Features

Model-Theory of Property Grammars with Features Model-Theory of Property Grammars with Features Denys Duchier Thi-Bich-Hanh Dao firstname.lastname@univ-orleans.fr Yannick Parmentier Abstract In this paper, we present a model-theoretic description of

More information

The Semantics of Definite DPs 1. b. Argument Position: (i) [ A politician ] arrived from Washington. (ii) Joe likes [ the politician ].

The Semantics of Definite DPs 1. b. Argument Position: (i) [ A politician ] arrived from Washington. (ii) Joe likes [ the politician ]. The Semantics of Definite DPs 1 Thus far, our semantics is able to interpret common nouns that occupy predicate position (1a). However, the most common position for common nouns to occupy is internal to

More information

CAS LX 500 Topics in Linguistics: Questions Spring 2006 March 2, b: Prosody and Japanese wh-questions

CAS LX 500 Topics in Linguistics: Questions Spring 2006 March 2, b: Prosody and Japanese wh-questions CAS LX 500 Topics in Linguistics: Questions Spring 2006 March 2, 2006 Paul Hagstrom 7b: Prosody and Japanese wh-questions Prosody by phase Prosody and syntax seem to fit quite tightly together. Ishihara,

More information

Turing machines and linear bounded automata

Turing machines and linear bounded automata and linear bounded automata Informatics 2A: Lecture 29 John Longley School of Informatics University of Edinburgh jrl@inf.ed.ac.uk 27 November 2015 1 / 15 The Chomsky hierarchy: summary Level Language

More information

Other types of Movement

Other types of Movement Other types of Movement So far we seen Wh-movement, which moves certain types of (XP) constituents to the specifier of a CP. Wh-movement is also called A-bar movement. We will look at two more types of

More information

Spring 2017 Ling 620. An Introduction to the Semantics of Tense 1

Spring 2017 Ling 620. An Introduction to the Semantics of Tense 1 1. Introducing Evaluation Times An Introduction to the Semantics of Tense 1 (1) Obvious, Fundamental Fact about Sentences of English The truth of some sentences (of English) depends upon the time they

More information

Quantification: Quantifiers and the Rest of the Sentence

Quantification: Quantifiers and the Rest of the Sentence Ling255: Sem & Cogsci Maribel Romero February 17, 2005 Quantification: Quantifiers and the Rest of the Sentence 1. Introduction. We have seen that Determiners express a relation between two sets of individuals

More information

CAS LX 522 Syntax I. We give trees to ditransitives. We give trees to ditransitives. We give trees to ditransitives. Problems continue * VP

CAS LX 522 Syntax I. We give trees to ditransitives. We give trees to ditransitives. We give trees to ditransitives. Problems continue * VP 9 CAS LX 522 Syntax I UTAH (4.3-4.4) You may recall our discussion of θ-theory, where we triumphantly classified erbs as coming in (at least) three types: Intransitie (1 θ-role) Transitie (2 θ-roles) Ditransitie

More information

THE DRAVIDIAN EXPERIENCER CONSTRUCTION AND THE ENGLISH SEEM CONSTRUCTION. K. A. Jayaseelan CIEFL, Hyderabad

THE DRAVIDIAN EXPERIENCER CONSTRUCTION AND THE ENGLISH SEEM CONSTRUCTION. K. A. Jayaseelan CIEFL, Hyderabad THE DRAVIDIAN EXPERIENCER CONSTRUCTION AND THE ENGLISH SEEM CONSTRUCTION K. A. Jayaseelan CIEFL, Hyderabad 1. Introduction In many languages e.g. Malayalam, Tamil, Hindi, the same verb is used in the Experiencer

More information

Linearization. {α,β}= def. α and β aresisters

Linearization. {α,β}= def. α and β aresisters Linearization Linguistics 601 3 ecember 2015 Our starting position in thinking about linearization was built upon a view of phrase structure that distinguished maximal projections from intermediary phrases.

More information

CS 3110: Proof Strategy and Examples. 1 Propositional Logic Proof Strategy. 2 A Proof Walkthrough

CS 3110: Proof Strategy and Examples. 1 Propositional Logic Proof Strategy. 2 A Proof Walkthrough CS 3110: Proof Strategy and Examples 1 Propositional Logic Proof Strategy The fundamental thing you have to do is figure out where each connective is going to come from. Sometimes the answer is very simple;

More information

BARE PHRASE STRUCTURE, LCA AND LINEARIZATION

BARE PHRASE STRUCTURE, LCA AND LINEARIZATION Neuere Entwicklungen der Grammatiktheorie Universität Wien, SoSe 2008 Winfried Lechner wlechner@gs.uoa.gr Handout #2 BARE PHRASE STRUCTURE, LCA AND LINEARIZATION http://vivaldi.sfs.nphil.uni-tuebingen.de/%7ennsle01/synvie.htm

More information

Dependency grammar. Recurrent neural networks. Transition-based neural parsing. Word representations. Informs Models

Dependency grammar. Recurrent neural networks. Transition-based neural parsing. Word representations. Informs Models Dependency grammar Morphology Word order Transition-based neural parsing Word representations Recurrent neural networks Informs Models Dependency grammar Morphology Word order Transition-based neural parsing

More information

Computational Linguistics

Computational Linguistics Computational Linguistics Dependency-based Parsing Clayton Greenberg Stefan Thater FR 4.7 Allgemeine Linguistik (Computerlinguistik) Universität des Saarlandes Summer 2016 Acknowledgements These slides

More information

Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing

Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing Context-Free Parsing: CKY & Earley Algorithms and Probabilistic Parsing Natural Language Processing CS 4120/6120 Spring 2017 Northeastern University David Smith with some slides from Jason Eisner & Andrew

More information

Generalized Quantifiers Logical and Linguistic Aspects

Generalized Quantifiers Logical and Linguistic Aspects Generalized Quantifiers Logical and Linguistic Aspects Lecture 1: Formal Semantics and Generalized Quantifiers Dag Westerståhl University of Gothenburg SELLC 2010 Institute for Logic and Cognition, Sun

More information

Roger Levy Probabilistic Models in the Study of Language draft, October 2,

Roger Levy Probabilistic Models in the Study of Language draft, October 2, Roger Levy Probabilistic Models in the Study of Language draft, October 2, 2012 224 Chapter 10 Probabilistic Grammars 10.1 Outline HMMs PCFGs ptsgs and ptags Highlight: Zuidema et al., 2008, CogSci; Cohn

More information

0A B 0C 0D 0E 0F A 0B 0C 0D 0E 0F

0A B 0C 0D 0E 0F A 0B 0C 0D 0E 0F 0A 0 03 06 0 0 0 01 02 03 04 05 06 07 08 09 0B 0C 0D 0E 0F 0 0 0 0 0 0 0 01 02 03 04 05 06 07 08 09 0A 0B 0C 0D 0E 0F 0 0 0 0 0 0 1 kracht.tex; 9/02/2005; 16:01; no v.; p.2 Syntactic Codes and Grammar

More information

Spring 2018 Ling 620 The Basics of Intensional Semantics, Part 1: The Motivation for Intensions and How to Formalize Them 1

Spring 2018 Ling 620 The Basics of Intensional Semantics, Part 1: The Motivation for Intensions and How to Formalize Them 1 The Basics of Intensional Semantics, Part 1: The Motivation for Intensions and How to Formalize Them 1 1. The Inadequacies of a Purely Extensional Semantics (1) Extensional Semantics a. The interpretation

More information

Categorial Grammar. Larry Moss NASSLLI. Indiana University

Categorial Grammar. Larry Moss NASSLLI. Indiana University 1/37 Categorial Grammar Larry Moss Indiana University NASSLLI 2/37 Categorial Grammar (CG) CG is the tradition in grammar that is closest to the work that we ll do in this course. Reason: In CG, syntax

More information

HPSG II: the plot thickens

HPSG II: the plot thickens Syntactic Models 2/21/06 HPSG II: the plot thickens 1 Passive: a lexical rule that rearranges ARG-ST! (1) Passive Lexical Rule < 1, tv - lxm ARG - ST INDEX i < FPSP 1 a, > part - lxm SYN HEAD FORM pass

More information

Computational Linguistics. Acknowledgements. Phrase-Structure Trees. Dependency-based Parsing

Computational Linguistics. Acknowledgements. Phrase-Structure Trees. Dependency-based Parsing Computational Linguistics Dependency-based Parsing Dietrich Klakow & Stefan Thater FR 4.7 Allgemeine Linguistik (Computerlinguistik) Universität des Saarlandes Summer 2013 Acknowledgements These slides

More information

Movement-Generalized Minimalist Grammars

Movement-Generalized Minimalist Grammars Movet-Generalized Minimalist Grammars Thomas Graf tgraf@ucla.edu tgraf.bol.ucla.edu University of California, Los Angeles LACL 2012 July 2, 2012 Outline 1 Monadic Second-Order Logic (MSO) Talking About

More information

Semantics and Pragmatics of NLP

Semantics and Pragmatics of NLP Semantics and Pragmatics of NLP Alex Ewan School of Informatics University of Edinburgh 28 January 2008 1 2 3 Taking Stock We have: Introduced syntax and semantics for FOL plus lambdas. Represented FOL

More information

2013 ISSN: JATLaC Journal 8: t 1. t t Chomsky 1993 I Radford (2009) R I t t R I 2. t R t (1) (= R's (15), p. 86) He could have helped

2013 ISSN: JATLaC Journal 8: t 1. t t Chomsky 1993 I Radford (2009) R I t t R I 2. t R t (1) (= R's (15), p. 86) He could have helped t 1. tt Chomsky 1993 IRadford (2009) R Itt R I 2. t R t (1) (= R's (15), p. 86) He could have helped her, or [she have helped him]. 2 have I has had I I could ellipsis I gapping R (2) (= R (18), p.88)

More information

Semantics and Generative Grammar. Quantificational DPs, Part 3: Covert Movement vs. Type Shifting 1

Semantics and Generative Grammar. Quantificational DPs, Part 3: Covert Movement vs. Type Shifting 1 Quantificational DPs, Part 3: Covert Movement vs. Type Shifting 1 1. Introduction Thus far, we ve considered two competing analyses of sentences like those in (1). (1) Sentences Where a Quantificational

More information

Vectors for Zero Robotics students

Vectors for Zero Robotics students Vectors for Zero Robotics students Zero Robotics Australia August 7, 08 Assumed Knowledge The robots used for the Zero Robotics competition (SPHERES) were designed for NASA researchers, and are able to

More information

Presuppositions (introductory comments)

Presuppositions (introductory comments) 1 Presuppositions (introductory comments) Some examples (1) a. The person who broke the typewriter was Sam. b. It was Sam who broke the typewriter. c. John screwed up again. d. John likes Mary, too. e.

More information

Maschinelle Sprachverarbeitung

Maschinelle Sprachverarbeitung Maschinelle Sprachverarbeitung Parsing with Probabilistic Context-Free Grammar Ulf Leser Content of this Lecture Phrase-Structure Parse Trees Probabilistic Context-Free Grammars Parsing with PCFG Other

More information

Maschinelle Sprachverarbeitung

Maschinelle Sprachverarbeitung Maschinelle Sprachverarbeitung Parsing with Probabilistic Context-Free Grammar Ulf Leser Content of this Lecture Phrase-Structure Parse Trees Probabilistic Context-Free Grammars Parsing with PCFG Other

More information

CS1800: Strong Induction. Professor Kevin Gold

CS1800: Strong Induction. Professor Kevin Gold CS1800: Strong Induction Professor Kevin Gold Mini-Primer/Refresher on Unrelated Topic: Limits This is meant to be a problem about reasoning about quantifiers, with a little practice of other skills, too

More information

Parsing Beyond Context-Free Grammars: Tree Adjoining Grammars

Parsing Beyond Context-Free Grammars: Tree Adjoining Grammars Parsing Beyond Context-Free Grammars: Tree Adjoining Grammars Laura Kallmeyer & Tatiana Bladier Heinrich-Heine-Universität Düsseldorf Sommersemester 2018 Kallmeyer, Bladier SS 2018 Parsing Beyond CFG:

More information

One hint from secondary predication (from Baker 1997 (8) A secondary predicate cannot take the goal argument as subject of predication, wheth

One hint from secondary predication (from Baker 1997 (8) A secondary predicate cannot take the goal argument as subject of predication, wheth MIT, Fall 2003 1 The Double Object Construction (Larson 1988, Aoun & Li 1989) MIT, 24.951, Fr 14 Nov 2003 A familiar puzzle The Dative Alternation (1) a. I gave the candy to the children b. I gave the

More information

Recurrent neural network grammars

Recurrent neural network grammars Widespread phenomenon: Polarity items can only appear in certain contexts Recurrent neural network grammars lide credits: Chris Dyer, Adhiguna Kuncoro Example: anybody is a polarity item that tends to

More information

Probabilistic Context-free Grammars

Probabilistic Context-free Grammars Probabilistic Context-free Grammars Computational Linguistics Alexander Koller 24 November 2017 The CKY Recognizer S NP VP NP Det N VP V NP V ate NP John Det a N sandwich i = 1 2 3 4 k = 2 3 4 5 S NP John

More information

Lecture 13: Soundness, Completeness and Compactness

Lecture 13: Soundness, Completeness and Compactness Discrete Mathematics (II) Spring 2017 Lecture 13: Soundness, Completeness and Compactness Lecturer: Yi Li 1 Overview In this lecture, we will prvoe the soundness and completeness of tableau proof system,

More information

Sentence Planning 2: Aggregation

Sentence Planning 2: Aggregation Pipelined Microplanning Sentence Planning 2: Aggregation Lecture 10 February 28, 2012 Reading: Chapter 5, Reiter and Dale Document Plan! Lexical choice! PPSs Aggregation PPSs Referring Expression Gen!

More information

Manual of Logical Style

Manual of Logical Style Manual of Logical Style Dr. Holmes January 9, 2015 Contents 1 Introduction 2 2 Conjunction 3 2.1 Proving a conjunction...................... 3 2.2 Using a conjunction........................ 3 3 Implication

More information

Wh-movement. CAS LX 522 Syntax I Fall 2001 November 6, 2001

Wh-movement. CAS LX 522 Syntax I Fall 2001 November 6, 2001 AS LX 522 Syntax I Fall 200 November 6, 200 Paul Hagstrom Week 9: Wh-movement Preliminary tree to remind ourselves: They will bake a cake. () P [ Q] T T VP will P bake they NP a N cake Verb bake needs

More information

A Multi-Modal Combinatory Categorial Grammar. A MMCCG Analysis of Japanese Nonconstituent Clefting

A Multi-Modal Combinatory Categorial Grammar. A MMCCG Analysis of Japanese Nonconstituent Clefting A Multi-Modal Combinatory Categorial Grammar Analysis of Japanese Nonconstituent Clefting Department of Linguistics The Ohio State University {kubota,esmith}@ling.osu.edu A basic cleft sentence Data Clefting

More information

ALGEBRA. 1. Some elementary number theory 1.1. Primes and divisibility. We denote the collection of integers

ALGEBRA. 1. Some elementary number theory 1.1. Primes and divisibility. We denote the collection of integers ALGEBRA CHRISTIAN REMLING 1. Some elementary number theory 1.1. Primes and divisibility. We denote the collection of integers by Z = {..., 2, 1, 0, 1,...}. Given a, b Z, we write a b if b = ac for some

More information

Semantics and Generative Grammar. An Introduction to Intensional Semantics 1

Semantics and Generative Grammar. An Introduction to Intensional Semantics 1 An Introduction to Intensional Semantics 1 1. The Inadequacies of a Purely Extensional Semantics (1) Our Current System: A Purely Extensional Semantics The extension of a complex phrase is (always) derived

More information

Introduction to Computational Linguistics

Introduction to Computational Linguistics Introduction to Computational Linguistics Olga Zamaraeva (2018) Based on Bender (prev. years) University of Washington May 3, 2018 1 / 101 Midterm Project Milestone 2: due Friday Assgnments 4& 5 due dates

More information

Lecture 10: Gentzen Systems to Refinement Logic CS 4860 Spring 2009 Thursday, February 19, 2009

Lecture 10: Gentzen Systems to Refinement Logic CS 4860 Spring 2009 Thursday, February 19, 2009 Applied Logic Lecture 10: Gentzen Systems to Refinement Logic CS 4860 Spring 2009 Thursday, February 19, 2009 Last Tuesday we have looked into Gentzen systems as an alternative proof calculus, which focuses

More information

AP Physics 1 Summer Assignment

AP Physics 1 Summer Assignment Name: Email address (write legibly): AP Physics 1 Summer Assignment Packet 3 The assignments included here are to be brought to the first day of class to be submitted. They are: Problems from Conceptual

More information

Turing machines and linear bounded automata

Turing machines and linear bounded automata and linear bounded automata Informatics 2A: Lecture 30 John Longley School of Informatics University of Edinburgh jrl@inf.ed.ac.uk 25 November 2016 1 / 17 The Chomsky hierarchy: summary Level Language

More information

Semantics and Generative Grammar. The Semantics of Adjectival Modification 1. (1) Our Current Assumptions Regarding Adjectives and Common Ns

Semantics and Generative Grammar. The Semantics of Adjectival Modification 1. (1) Our Current Assumptions Regarding Adjectives and Common Ns The Semantics of Adjectival Modification 1 (1) Our Current Assumptions Regarding Adjectives and Common Ns a. Both adjectives and common nouns denote functions of type (i) [[ male ]] = [ λx : x D

More information