The SUBTLE NL Parsing Pipeline: A Complete Parser for English Mitch Marcus University of Pennsylvania

Similar documents
A Context-Free Grammar

CS 545 Lecture XVI: Parsing

Penn Treebank Parsing. Advanced Topics in Language Processing Stephen Clark

Probabilistic Context-Free Grammars. Michael Collins, Columbia University

Probabilistic Context Free Grammars. Many slides from Michael Collins

Probabilistic Context-free Grammars

Probabilistic Context Free Grammars. Many slides from Michael Collins and Chris Manning

Recap: Lexicalized PCFGs (Fall 2007): Lecture 5 Parsing and Syntax III. Recap: Charniak s Model. Recap: Adding Head Words/Tags to Trees

LECTURER: BURCU CAN Spring

Constituency Parsing

Parsing with Context-Free Grammars

Terry Gaasterland Scripps Institution of Oceanography University of California San Diego, La Jolla, CA,

CS 6120/CS4120: Natural Language Processing

S NP VP 0.9 S VP 0.1 VP V NP 0.5 VP V 0.1 VP V PP 0.1 NP NP NP 0.1 NP NP PP 0.2 NP N 0.7 PP P NP 1.0 VP NP PP 1.0. N people 0.

Multilevel Coarse-to-Fine PCFG Parsing

Natural Language Processing CS Lecture 06. Razvan C. Bunescu School of Electrical Engineering and Computer Science

Maschinelle Sprachverarbeitung

Maschinelle Sprachverarbeitung

Computational Linguistics

Statistical Methods for NLP

Computational Linguistics. Acknowledgements. Phrase-Structure Trees. Dependency-based Parsing

Multiword Expression Identification with Tree Substitution Grammars

Transition-Based Parsing

Natural Language Processing

13A. Computational Linguistics. 13A. Log-Likelihood Dependency Parsing. CSC 2501 / 485 Fall 2017

Part of Speech Tagging: Viterbi, Forward, Backward, Forward- Backward, Baum-Welch. COMP-599 Oct 1, 2015

NLP Homework: Dependency Parsing with Feed-Forward Neural Network

Introduction to the CoNLL-2004 Shared Task: Semantic Role Labeling

10/17/04. Today s Main Points

Parsing. Based on presentations from Chris Manning s course on Statistical Parsing (Stanford)

The Infinite PCFG using Hierarchical Dirichlet Processes

Part-of-Speech Tagging + Neural Networks CS 287

The Research on Syntactic Features in Semantic Role Labeling

Advanced Natural Language Processing Syntactic Parsing

TALP at GeoQuery 2007: Linguistic and Geographical Analysis for Query Parsing

Lecture 13: Structured Prediction

Personal Project: Shift-Reduce Dependency Parsing

Features of Statistical Parsers

Natural Language Processing 1. lecture 7: constituent parsing. Ivan Titov. Institute for Logic, Language and Computation

Parsing with Context-Free Grammars

NLP Programming Tutorial 11 - The Structured Perceptron

A Deterministic Word Dependency Analyzer Enhanced With Preference Learning

Latent Variable Models in NLP

Statistical methods in NLP, lecture 7 Tagging and parsing

HMM and Part of Speech Tagging. Adam Meyers New York University

The Rise of Statistical Parsing

Natural Language Processing

Structured language modeling

INF4820: Algorithms for Artificial Intelligence and Natural Language Processing. Hidden Markov Models

A Syntax-based Statistical Machine Translation Model. Alexander Friedl, Georg Teichtmeister

Hidden Markov Models (HMMs)

Introduction to Semantic Parsing with CCG

Marrying Dynamic Programming with Recurrent Neural Networks

c(a) = X c(a! Ø) (13.1) c(a! Ø) ˆP(A! Ø A) = c(a)

Posterior vs. Parameter Sparsity in Latent Variable Models Supplementary Material

Processing/Speech, NLP and the Web

Transformational Priors Over Grammars

Introduction to Data-Driven Dependency Parsing

Natural Language Processing : Probabilistic Context Free Grammars. Updated 5/09

Structured Prediction

INF4820: Algorithms for Artificial Intelligence and Natural Language Processing. Hidden Markov Models

Language Processing with Perl and Prolog

Boosting Applied to Tagging and PP Attachment

Natural Language Processing

Spectral Unsupervised Parsing with Additive Tree Metrics

Alessandro Mazzei MASTER DI SCIENZE COGNITIVE GENOVA 2005

Driving Semantic Parsing from the World s Response

Lecture 7: Introduction to syntax-based MT

Lab 12: Structured Prediction

Multi-Component Word Sense Disambiguation

Dependency grammar. Recurrent neural networks. Transition-based neural parsing. Word representations. Informs Models

Extraction of Opposite Sentiments in Classified Free Format Text Reviews

Lecture 11: PCFGs: getting luckier all the time

Empirical Methods in Natural Language Processing Lecture 11 Part-of-speech tagging and HMMs

Hidden Markov Models

Chapter 14 (Partially) Unsupervised Parsing

Spatial Role Labeling: Towards Extraction of Spatial Relations from Natural Language

Spatial Role Labeling CS365 Course Project

Lecture 9: Hidden Markov Model

Aspects of Tree-Based Statistical Machine Translation

CS838-1 Advanced NLP: Hidden Markov Models

Syntax-based Statistical Machine Translation

Unit 2: Tree Models. CS 562: Empirical Methods in Natural Language Processing. Lectures 19-23: Context-Free Grammars and Parsing

A Supertag-Context Model for Weakly-Supervised CCG Parser Learning

Stochastic Parsing. Roberto Basili


Applied Natural Language Processing

AN ABSTRACT OF THE DISSERTATION OF

ACS Introduction to NLP Lecture 2: Part of Speech (POS) Tagging

Part-of-Speech Tagging

Chunking with Support Vector Machines

Bringing machine learning & compositional semantics together: central concepts

A Probabilistic Forest-to-String Model for Language Generation from Typed Lambda Calculus Expressions

Extracting Information from Text

Semantics and Generative Grammar. A Little Bit on Adverbs and Events

Regularization Introduction to Machine Learning. Matt Gormley Lecture 10 Feb. 19, 2018

A DOP Model for LFG. Rens Bod and Ronald Kaplan. Kathrin Spreyer Data-Oriented Parsing, 14 June 2005

Part I - Introduction Part II - Rule Extraction Part III - Decoding Part IV - Extensions

Low-Dimensional Discriminative Reranking. Jagadeesh Jagarlamudi and Hal Daume III University of Maryland, College Park

CS460/626 : Natural Language

Transcription:

The SUBTLE NL Parsing Pipeline: A Complete Parser for English Mitch Marcus University of Pennsylvania 1

PICTURE OF ANALYSIS PIPELINE Tokenize Maximum Entropy POS tagger MXPOST Ratnaparkhi Core Parser Collins Generative Prob. CFG Functional Tag Replacement Kulick Null Element Placement Ryan Gabbard PhD Dec 2010 (just submitted) 2

NLP Analysis Pipeline Typed text in Tokenize Maximum Entropy POS Tagger (Ratnaparkhi 97) Parse (Collins Bikel 04) Junior, defuse any bombs immediately when going into a room. Junior_NNP,_, defuse_vbp any_dt bombs_nns immediately_rb when_wrb going_vbg into_in a_dt room_nn._. ( (S (NP-VOC (NNP Junior)) (,,) (NP-SBJ-0 (-NONE- *)) (VP (VB defuse) (NP (DT any) (NNS bombs)) (ADVP-TMP (RB immediately)) (SBAR-TMP (WHADVP-1 (WRB when)) (..))) (S (NP-SBJ-0 (-NONE- *)) (VP (VBG going) (PP-DIR (IN into) (NP (DT a) (NN room))) (ADVP-1 (-NONE- *T*)))))) 11/4/2010 SUBTLE Annual Review 3

NLP Analysis Pipeline Typed text in Tokenize Maximum Entropy POS Tagger (Ratnaparkhi 97) Parse (Collins Bikel 04) Restore Function Tags (Kulick 06) Junior, defuse any bombs immediately when going into a room. Junior_NNP,_, defuse_vbp any_dt bombs_nns immediately_rb when_wrb going_vbg into_in a_dt room_nn._. ( (S (NP-VOC (NNP Junior)) (,,) (NP-SBJ-0 (-NONE- *)) (VP (VB defuse) (NP (DT any) (NNS bombs)) (ADVP-TMP (RB immediately)) (SBAR-TMP (WHADVP-1 (WRB when)) (..))) (S (NP-SBJ-0 (-NONE- *)) (VP (VBG going) (PP-DIR (IN into) (NP (DT a) (NN room))) (ADVP-1 (-NONE- *T*)))))) 11/4/2010 SUBTLE Annual Review 4

NLP Analysis Pipeline Typed text in Tokenize Maximum Entropy POS Tagger (Ratnaparkhi 97) Parse (Collins Bikel 04) Restore Function Tags (Kulick 06) Restore Null Elements (Gabbard 10) Junior, defuse any bombs immediately when going into a room. Junior_NNP,_, defuse_vbp any_dt bombs_nns immediately_rb when_wrb going_vbg into_in a_dt room_nn._. ( (S (NP-VOC (NNP Junior)) (,,) (NP-SBJ-0 (-NONE- *)) (VP (VB defuse) (NP (DT any) (NNS bombs)) (ADVP-TMP (RB immediately)) (SBAR-TMP (WHADVP-1 (WRB when)) (..))) (S (NP-SBJ-0 (-NONE- *)) (VP (VBG going) (PP-DIR (IN into) (NP (DT a) (NN room))) (ADVP-1 (-NONE- *T*)))))) 11/4/2010 SUBTLE Annual Review 5

NLP Analysis Pipeline 11/4/2010 Typed text in Tokenize Maximum Entropy POS Tagger (Ratnaparkhi 97) Parse (Collins Bikel 04) Restore Function Tags (Kulick 06) Restore Null Elements (Gabbard 10) Semantic Analysis (Perera) Junior, defuse any bombs immediately when going into a room. Junior_NNP,_, defuse_vbp any_dt bombs_nns immediately_rb when_wrb going_vbg into_in a_dt room_nn._. ( (S (NP-VOC (NNP Junior)) (,,) (NP-SBJ-0 (-NONE- *)) (VP (VB defuse) (NP (DT any) (NNS bombs)) (ADVP-TMP (RB immediately)) (SBAR-TMP (WHADVP-1 (WRB when)) (..))) (S (NP-SBJ-0 (-NONE- *)) (VP (VBG going) SUBTLE Annual Review (PP-DIR (IN into) (NP (DT a) (NN room))) (ADVP-1 (-NONE- *T*)))))) 6

1995: A breakthrough in parsing 10 6 words of Penn Treebank Annotation + Machine Learning = Robust Parsers training sentences (Magerman 95) Training Program answers The founder of Pakistan's nuclear program, Abdul Qadeer Khan, has admitted he transferred nuclear technology to Iran, Libya and North Korea 1990 Best hand-built parsers: 1995+ Statistical parsers: (both on short sentences) Models Parser ~40% accuracy >90% accuracy 7 S VP VP SBA R NP S VP NP PP NP PP NP NP NP NP NP NP NP NP The founder of Pakistan s nuclear department Abdul Qadeer Khan has admitted he transferred nuclear technology to Iran, Libya, and North Korea

Generative Parsing Results (<40 w) Parser F 1 Magerman 95 84.7 Collins 96 86.0 Charniak 97 87.4 Collins 99 (Bikel 04) 88.6 Liang 08 (extending Charniak 05) 91.7 8

Meaningless Progress.. 9

Treebank Tree with Function Tags NP-SBJ S VP Junior (S (NP-SBJ (NNP Junior)) (VP (MD should) (VP (VB defuse) (NP (NP (NNS bombs)) (PP-LOC (IN in) (NP (DT a) should (NN room)))) defuse (ADVP-TMP (RB immediately)))) VP NP NP any bombs PP-LOC in NP the room NP-TMP immediately 10

Treebank Tree with Null Elements SBARQ WHNP-1 Who was SQ VP NP-SBJ-2 *T*-1 S believed NP-SBJ-3 to have been *T*-2 VP VP shot NP believe(somebody, shoot (SOMEONE, Who)) *-3 11

A Perfect Parseval Analysis.. SBARQ WHNP SQ Who was believed VP S Without null element, can t tell what the semantic role of Who was, or even the predication it s part of. VP to have been VP shot believe(somebody, shoot (SOMEONE, Who)) 12

A Complete Parsing System Ryan Gabbard, Seth Kulick, and Mitch Marcus. Fully Parsing the Penn Treebank. HLT/NAACL 2006 Ryan Gabbard, Null Element Restoration, PhD Dissertation, Nov 2010 13

Step 1: Recovering Function Tags 14

Treebank II Function Tag Set Syntactic Semantic Miscellaneous SBJ Subject TMP Temporal CLF It-cleft LGS Logical Subject LOC Location HLN Headline PRD Predicate DIR Direction TTL Title DTV Dative EXT Extent Topicalization PUT Locative of put VOC Vocative MNR Manner BNF Benefactive TPC Topic PRP Purpose NOM Nominal CLR ADV Non-specific adverbial CLR Closely- Related 15

A Sample PCFG 16

An Example Derivation 17

The Key Trick in 95: Lexicalizing the CFG New problem: Estimating the probability of S(bought) NP(yesterday) NP(Marks) VP(bought) 18

Solving the Sparse Data Problem with Independence Assumptions Instead of S(bought) NP(yesterday) NP(Marks) VP(bought) 1. Generate Head, given lexicalized parent: S(bought) VP(bought). 2. Generate other children given parent & head: S(bought) NP( ) NP ( ) VP(bought) 3. Lexicalize other children, given parent, head & node type: S(bought) NP(yesterday)NP(Marks) VP(bought) 19

To augment the parser output with function tags 1. Find the code in Collins Model II that removes function tags 2. Remove it! {NP-OBJ( )} NP-OBJ( ) NP-TMP( ) NP-OBJ(her) NP-TMP(today) (From Collins & Koo 05) 20

Results for Function Tag Groups & Overall Overall w/ CLR Overall w/o CLR Collins 2 +FTags 88.95 90.78 Blaheta, 2003 88.28 Jijkoun & de Rijke, 2004 88.50 With no drop in basic parser F score! 21

Results for Function Tag Groups & Overall Overall w/ CLR Overall w/o CLR Syn 56% Sem 36% Collins 2 +FTags 88.95 90.78 95.76 84.56 Blaheta, 2003 88.28 Jijkoun & de Rijke, 2004 88.50 Musillo & Merlo, Oct 2005 96.5 85.6 With no drop in basic parser F score! 22

Step 2: Recovering Null Elements 23

Null Elements for Wh-movement SBARQ WHNP-1 SQ Predicate argument structure: see(your,what) What NP-SBJ VP did you see NP *T*-1 *T* - marks WH-questions Null elements are co-indexed using integer tags. To recover Pred-arg structure: Replace null element with co-indexed lexical material Similar approaches for other grammatical phenomenon 24

Previous Work Forward Integrated with Parsing (Collins Model 3, Dienes&Dubey 03) Postprocessing Parser Output Hand Built Rule-based (Campbell 04) Statistical Machine Learning Approaches (Johnson 02, Levy&Manning 04, Jijkoun&de Rijke 04) Our approach combines: Campbell s Problem Decomposition Machine Learning Into: A pipeline of linear classifiers 25

The Null Element Replacement Algorithm Pipeline & rich linguistic features á la Campbell Pipeline of linear classifiers (McCallum s Mallet) parser output with function tags NPTrace (NP*)insert NullComp ProAntecedent WHXPInsert-> WHXPDiscern Antecedentless WHTrace 26

Aggregate results over all recovered categories Automatically parsed trees from section 23 27

Present vs. Campbell s Rule Based System Test on gold--standard trees from section 23 using Campbell's metric 28

Wh placement with factor graphs 29

Wh placement with factor graphs 30

Results: two systems for Wh-placement 31

Summary: Full Parsing! Before S NP VP After S NP-SBJ-1 VP patients were VBD VP VBN examined patients VBD were 20 Function Tags Mostly syntactic or semantic Syntactic Tags: 95% recovery accuracy Semantic Tags: 84% recovery accuracy Traces Passivization, control, raising: ~70% recovered Wh-traces: ~80% recovered VP VBN examined NP-1 * 32