Alternative Syntactic Theories. L614 Spring 2010

Similar documents
Syntactic Theory. Background and Transformational Grammar. Dr. Dan Flickinger & PD Dr. Valia Kordoni

Outline of today s lecture

Pushdown automata. Informatics 2A: Lecture 9. Alex Simpson. 3 October, School of Informatics University of Edinburgh als@inf.ed.ac.

What s in a Lexicon. The Lexicon. Lexicon vs. Dictionary. What kind of Information should a Lexicon contain?

Noam Chomsky: Aspects of the Theory of Syntax notes

The Halting Problem is Undecidable

SYNTAX: THE ANALYSIS OF SENTENCE STRUCTURE

Learning Translation Rules from Bilingual English Filipino Corpus

Module Catalogue for the Bachelor Program in Computational Linguistics at the University of Heidelberg

Constraints in Phrase Structure Grammar

Paraphrasing controlled English texts

Phrase Structure Rules, Tree Rewriting, and other sources of Recursion Structure within the NP

6-1. Process Modeling

KSE Comp. support for the writing process 2 1

Managing Variability in Software Architectures 1 Felix Bachmann*

What is Organizational Communication?

MARY. V NP NP Subject Formation WANT BILL S

Mildly Context Sensitive Grammar Formalisms

3. Mathematical Induction

Chapter 13, Sections Auxiliary Verbs CSLI Publications

Basic Parsing Algorithms Chart Parsing

03 - Lexical Analysis

How the Computer Translates. Svetlana Sokolova President and CEO of PROMT, PhD.

Impossible Words? Jerry Fodor and Ernest Lepore

Ling 201 Syntax 1. Jirka Hana April 10, 2006

Introduction to Automata Theory. Reading: Chapter 1

Part 1 Expressions, Equations, and Inequalities: Simplifying and Solving

Automata and Computability. Solutions to Exercises

L130: Chapter 5d. Dr. Shannon Bischoff. Dr. Shannon Bischoff () L130: Chapter 5d 1 / 25

Artificial Intelligence Exam DT2001 / DT2006 Ordinarie tentamen

Overview of E0222: Automata and Computability

Semantic analysis of text and speech

Section 4.1 Rules of Exponents

Finite Automata and Formal Languages

Honors Class (Foundations of) Informatics. Tom Verhoeff. Department of Mathematics & Computer Science Software Engineering & Technology

Historical Linguistics. Diachronic Analysis. Two Approaches to the Study of Language. Kinds of Language Change. What is Historical Linguistics?

CS 6740 / INFO Ad-hoc IR. Graduate-level introduction to technologies for the computational treatment of information in humanlanguage

COMPUTER SCIENCE, BACHELOR OF SCIENCE (B.S.)

Fewer. Bigger. Stronger.

Teaching Formal Methods for Computational Linguistics at Uppsala University

Culture and Language. What We Say Influences What We Think, What We Feel and What We Believe

A chart generator for the Dutch Alpino grammar

Syntax: Phrases. 1. The phrase

Natural Language Database Interface for the Community Based Monitoring System *

A Chart Parsing implementation in Answer Set Programming

Organizational Issues Arising from the Integration of the Lexicon and Concept Network in a Text Understanding System

Introduction to Semantics. A Case Study in Semantic Fieldwork: Modality in Tlingit

No Solution Equations Let s look at the following equation: 2 +3=2 +7

PP-Attachment. Chunk/Shallow Parsing. Chunk Parsing. PP-Attachment. Recall the PP-Attachment Problem (demonstrated with XLE):

Math 4310 Handout - Quotient Vector Spaces

3. Logical Reasoning in Mathematics

Overview of MT techniques. Malek Boualem (FT)

Semantics and Generative Grammar. Quantificational DPs, Part 3: Covert Movement vs. Type Shifting 1

CINTIL-PropBank. CINTIL-PropBank Sub-corpus id Sentences Tokens Domain Sentences for regression atsts 779 5,654 Test

Motivation. Korpus-Abfrage: Werkzeuge und Sprachen. Overview. Languages of Corpus Query. SARA Query Possibilities 1

Engineering Process Software Qualities Software Architectural Design

The tiger quickly disappeared into the trees. The big cat vanished into the forest. Adolescent employees sometimes argue with their employers.

Computer Architecture Syllabus of Qualifying Examination

Introduction. Compiler Design CSE 504. Overview. Programming problems are easier to solve in high-level languages

Statistical Machine Translation

Formal Languages and Automata Theory - Regular Expressions and Finite Automata -

Symbiosis of Evolutionary Techniques and Statistical Natural Language Processing

Introduction to formal semantics -

Computer Science/Software Engineering

Compiler I: Syntax Analysis Human Thought

Computer Science 281 Binary and Hexadecimal Review

31 Case Studies: Java Natural Language Tools Available on the Web

Lecture 9. Phrases: Subject/Predicate. English 3318: Studies in English Grammar. Dr. Svetlana Nuernberg

Philosophical argument

CHECKLIST FOR THE DEGREE PROJECT REPORT

Modern foreign languages

According to the Argentine writer Jorge Luis Borges, in the Celestial Emporium of Benevolent Knowledge, animals are divided

Linguistics & Cognitive Science

Unit 2 Module 3: Generating Examples and Nonexamples

Constituency. The basic units of sentence structure

CS 533: Natural Language. Word Prediction

Task: Representing the National Debt 7 th grade

POLITE ENGLISH. Giving advice FREE ON-LINE COURSE. Lesson 2: version without a key SZKOLENIA JĘZYKOWE DLA FIRM ZREALIZUJEMY TWÓJ CEL!

TECH. Requirements. Why are requirements important? The Requirements Process REQUIREMENTS ELICITATION AND ANALYSIS. Requirements vs.

Open Domain Information Extraction. Günter Neumann, DFKI, 2012

CS4025: Pragmatics. Resolving referring Expressions Interpreting intention in dialogue Conversational Implicature

Guide to Writing the AP English Language Synthesis Essay. Wednesday, December 5, 12

Session 7 Fractions and Decimals

Three Ways to Clarify Your Writing

IP PATTERNS OF MOVEMENTS IN VSO TYPOLOGY: THE CASE OF ARABIC

Semantic Analysis: Types and Type Checking

Phase 2 of the D4 Project. Helmut Schmid and Sabine Schulte im Walde

Proposal Style Guide Quick Reference

6.080/6.089 GITCS Feb 12, Lecture 3

Open-Source, Cross-Platform Java Tools Working Together on a Dialogue System

Regions in a circle. 7 points 57 regions

A Beautiful Four Days in Berlin Takafumi Maekawa (Ryukoku University)

Overview of the TACITUS Project

Psychology G4470. Psychology and Neuropsychology of Language. Spring 2013.

Chapter Four: How to Collaborate and Write With Others

Transcription:

Alternative Syntactic Theories L614 Spring 2010

Syntactic analysis Generative grammar = collection of words and rules with which we generate strings of those words, i.e., sentences Syntax attempts to capture the nature of those rules (1) Colorless green ideas sleep furiously. (2) *Furiously sleep ideas green colorless. What generalizations are needed to capture the difference between grammatical sentences and ungrammatical sentences? 2/25

Syntax: What does it mean? Can view a syntactic theory in a number of ways, two of which are the following: Psychological model: syntactic structures correspond to what is in the heads of speakers Computational model: syntactic structures are formal objects which can be mathematically manipulated. We will focus on the computational way of viewing grammar for this class 3/25

Formalism vs. theory Will we actually look at theories?... Sort of. A theory describes a set of data and makes predictions for new data In this class, we will emphasize theories which are testable, i.e., can be verified or falsified A formalism provides a way of defining a theory with mathematical rigor It is essentially a set of beliefs and conditions that frame how generalizations can be made. The course name (Alternative Syntactic Theories) is a bit of a misnomer: we will actually be focusing on formalisms, and we will use theories to exemplify them. 4/25

The Transformational tradition Roughly speaking, transformational syntax (GB, P&P,...) has focused on the following: Explanatory adequacy: the data must fit with a deeper model, that of universal grammar Psychological: does the grammar make sense in light of what we know of how the mind works? Universality: generalizations must be applicable to all languages Transformations/Movement: (surface) sentences are derived from underlying other sentences, e.g. passives are derived from active sentences But this kind of theory often doesn t lend itself well to computational applications 5/25

Alternative assumptions Prioritize descriptive adequacy over explanatory adequacy Prioritize computational effectiveness over psychological reality e.g., movement is disfavored Prioritize description in one language before dealing with all languages The data will always be the same, but how you handle it, as we ll see, depends largely upon your assumptions 6/25

Making it computational How is a grammatical theory useful for computational lingusitics? Parsing: take an input sentence and return the syntactic analysis and/or state whether it is a valid sentence Generation: take a meaning representation and generate a valid sentence Both tasks are often subparts of practical applications (e.g., dialogue systems) 7/25

Computational needs To use a grammar for parsing or generation, we need to have a grammar that meets several criteria: Accurate: gives a correct analysis Precise: tells a computer exactly what it is that you want it to do Efficient: able to parse a sentence and return one or only a small number of parses Useful: is relatively easy to map a syntactic structure of a sentence to its meaning These needs are not necessarily why the computational formalisms were developed, but they are some of the reasons why people use them. 8/25

Computational Grammar Formalisms The formalisms we will look at this quarter generally share several properties: Descriptive adequacy Precise encodings (implementable) Constrained mathematical formalisms Monostratal frameworks (Usually) highly lexical 9/25

Descriptive adequacy Some researchers try to explain the underlying mechanisms, but we are most concerned with being able to describe linguistic phenomena Provide a structural description for every well-formed sentence Define which sentences are well-formed and which are not in a language Give us an accurate encoding of a language Interested in broad-coverage, i.e., can (try to) describe all of a language less of a distinction between core and periphery phenomena 10/25

Precise encodings Mathematical formalism: formal way to generate sets of strings Precisely define: elementary structures ways of combining those structures Such an emphasis on mathematical precision makes these grammar formalisms more easily implementable Will 2 parts of your grammar conflict? If we have precisely encoded the grammar, we can answer this question with certainty. 11/25

Constrained mathematical formalisms Formalism should (arguably) be constrained, i.e., cannot be allowed to specify all strings Linguistic motivation: Limits the scope of the theory of grammar Computational motivation: Allows us to define efficient processing models This is different than constraining a theory What is the minimum amount of mathematical overhead that we need to describe language? 12/25

Monostratal frameworks Only have one (surface) syntactic level Make no recourse to movement or transformations Augment your basic (phrase structure) tree with information that can describe movement phenomena Need some way to relate different structures (e.g., active and passive) without invoking, e.g., traces Without having to refer to movement, easier to process sentences on a computer. 13/25

Lexical In the past, rules applied to broad classes and only some information was put in the lexicon, e.g., subcategorization information. But more and more theories emphasize the role of individual lexical items in grammatical constructions Linguistic motivation: lexicon best way to specify some generalizations: He told/*divulged me the truth Computational motivation: can derive lexical information from corpora Shift more of the information to the lexicon; each lexical item may be a complex object. 14/25

Brief mention of complexity We have touched on the complexity of different formalisms Type Automaton Grammar Memory Name Rule Name 0 Unbounded TM α β General rewrite 1 Bounded LBA β A γ β δ γ Context-sensitive 2 Stack PDA A β Context-free 3 None FSA A xb, A x Right linear TM: Turing Machine LBA: Linear-Bounded Automaton PDA: Push-Down Automaton FSA: Finite-State Automaton 15/25

Criteria under which to evaluate grammar formalisms There are three kinds of criteria: linguistic naturalness mathematical power computational effectiveness and efficiency The weaker the type of grammar: the stronger the claim made about possible languages the greater the potential efficiency of the parsing procedure Reasons for choosing a stronger grammar class: to capture the empirical reality of actual languages to provide for elegant analyses capturing more generalizations ( more compact grammars) 16/25

Context-Free Grammars (CFGs) Context-Free Grammars (CFGs) are one kind of constrained mathematical formalism and probably the most popular for writing English grammars elementary structures: rules composed of nonterminal and terminal elements combine rules by rewriting them Example of a set of rules: S NP VP NP Det N VP V NP... But these rules are rather impoverished... 17/25

Are CFGs good enough? Data from Swiss German and other languages show that CFGs are not powerful enough to handle all natural language constructions CFGs are not easily lexicalized (and we need lexical knowledge) CFGs become complicated once we start taking into account agreement features, verb subcategorizations, unbounded dependency constructions, raising constructions, etc. We need more refined formalisms... 18/25

Beyond CFGs We want to move beyond CFGs to better capture language, but maintain that level of precision We can look at it a couple of ways: Extend the basic model of CFGS with, e.g., complex categories, functional structure, feature structures,... Eliminate CFG model (or derive it some other way) The frameworks we will investigate take one of these approaches 19/25

Computational Grammar Frameworks What we will look at the rest of the semester: Dependency Grammar (DG) Tree-Adjoining Grammar (TAG) Lexical-Functional Grammar (LFG) Head-driven Phrase Structure Grammar (HPSG) Combinatory Categorial Grammar (CCG) 20/25

Dependency Grammar (DG) The way to analyze a sentence is by looking at the relations between words No grouping, or constituency, is used DG traditions are often completely independent of constituency-based traditions (e.g., CFGs) DG is not a unified framework; there are a host of different frameworks within this tradition A verb and its arguments drive an analysis, which is closely related to the semantics of a sentence Some of the other frameworks we ll investigate utilize insights from DG 21/25

Tree-Adjoining Grammar (TAG) The analysis looks like a CFG tree, but the way to get it is completely different... Elementary structures are trees of arbitrary height Trees are rooted in lexical items, i.e. lexicalized In other words, the lexicon contains tree fragments as parts of lexical entries Put trees together by substituting and adjoining them, resulting in a final tree which looks like a CFG-derived tree 22/25

Lexical-Functional Grammar (LFG) Functional structure (subject, object, etc.) divided from constituent structure (tree structure) Kind of like combining dependency structure with phrase structure The f-structures are potentially very complex, however. Can express some generalizations in f-structure; some in c-structure; i.e., not restricted to saying everything in terms of trees 23/25

Head-driven Phrase Structure Grammar (HPSG) Sentences, phrases, and words all uniformly treated as linguistic signs, i.e., complex objects of features Many analyses rely on a CFG backbone, but this need not be so. Similar to LFG in its use of a feature architecture Uses an inheritance hierarchy to relate different types of objects (e.g., nouns and determiners are both types of nominals) 24/25

Combinatory Categorial Grammar (CCG) Categorial Grammar derives sentences in a proof-solving manner, maintaining a close link with a semantic representation Lexical categories specify how to combine words into sentences The idea of selection is crucial, e.g., a verb will select for the number and type of arguments Again, lexical entries contain tree-like information CCG has sophisticated mechnisms that deal nicely with coordination, extraction, and other constructions 25/25