INF4820: Algorithms for Artificial Intelligence and Natural Language Processing. Context-Free Grammars & Parsing
|
|
- Ann Cameron
- 7 years ago
- Views:
Transcription
1 University of Oslo : Department of Informatics INF4820: Algorithms for Artificial Intelligence and Natural Language Processing Context-Free Grammars & Parsing Stephan Oepen & Murhaf Fares Language Technology Group (LTG) November 3, 2016
2 Overview Last Time Sequence Labeling Dynamic programming Viterbi algorithm Forward algorithm
3 Overview Last Time Sequence Labeling Dynamic programming Viterbi algorithm Forward algorithm Today Grammatical structure Context-free grammar Treebanks Probabilistic CFGs
4 Recall: Ice Cream and Global Warming 0.8 S H 0.2 C 0.5 P(1 H)=0.2 P(2 H)=0.4 P(3 H)= /S 0.2 P(1 C) = 0.5 P(2 C) = 0.4 P(3 C) = 0.1
5 Recall: Viterbi Algorithm To find the best state sequence, maximize: P(s 1... s n o 1... o n ) = P(s 1 s 0 )P(o 1 s 1 )P(s 2 s 1 )P(o 2 s 2 )... The value we cache at each step: v i (s) = L max k=1 [v i 1(k) P(s k) P(o i s)] The variable v i (s) represents the maximum probability that the i-th state is s, given that we have seen O i 1. At each step, we record backpointers showing which previous state led to the maximum probability.
6 Recall: Dynamic Programming Dynamic programming algorithms solve large problems by compounding answers from smaller sub-problems record sub-problem solutions for repeated use
7 Recall: Dynamic Programming Dynamic programming algorithms solve large problems by compounding answers from smaller sub-problems record sub-problem solutions for repeated use They are used for complex problems that can be described recursively require the same calculations over and over again
8 Recall: Dynamic Programming Dynamic programming algorithms solve large problems by compounding answers from smaller sub-problems record sub-problem solutions for repeated use They are used for complex problems that can be described recursively require the same calculations over and over again Examples: Dijkstra s shortest path minimum edit distance longest common subsequence Viterbi decoding
9 Recall: An Example of the Viterbi Algorithmn H H H S /S C C C o 1 o 2 o 3
10 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 S P(H S)P(3 H) H H H /S C C C o 1 o 2 o 3
11 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 S P(H S)P(3 H) H H H /S C C C o 1 o 2 o 3
12 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 S P(H S)P(3 H) H H H /S P(C S)P(3 C) C C C v 1 (C) = o 1 o 2 o 3
13 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 S P(H S)P(3 H) H H H /S P(C S)P(3 C) C C C v 1 (C) = o 1 o 2 o 3
14 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 v 2 (H) = max(.32.12,.02.06) =.0384 S P(H S)P(3 H) P(C S)P(3 C) P(H H)P(1 H) H H H P(H C)P(1 H) C C C /S v 1 (C) = o 1 o 2 o 3
15 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 v 2 (H) = max(.32.12,.02.06) =.0384 S P(H S)P(3 H) P(C S)P(3 C) P(H H)P(1 H) H H H P(H C)P(1 H) C C C /S v 1 (C) = o 1 o 2 o 3
16 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 v 2 (H) = max(.32.12,.02.06) =.0384 S P(H S)P(3 H) P(C S)P(3 C) P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) C C C /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) = o 1 o 2 o 3
17 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 v 2 (H) = max(.32.12,.02.06) =.0384 S P(H S)P(3 H) P(C S)P(3 C) P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) C C C /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) = o 1 o 2 o 3
18 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 v 2 (H) = max(.32.12,.02.06) =.0384 v 3 (H) = max( , ) = S P(H S)P(3 H) P(C S)P(3 C) P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) P(H H)P(3 H) P(H C)P(3 H) C C C /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) = o 1 o 2 o 3
19 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 v 2 (H) = max(.32.12,.02.06) =.0384 v 3 (H) = max( , ) = S P(H S)P(3 H) P(C S)P(3 C) P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) P(H H)P(3 H) P(H C)P(3 H) C C C /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) = o 1 o 2 o 3
20 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 v 2 (H) = max(.32.12,.02.06) =.0384 v 3 (H) = max( , ) = S P(H S)P(3 H) P(C S)P(3 C) P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) P(H H)P(3 H) P(C H)P(3 C) P(H C)P(3 H) P(C C)P(3 C) C C C /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) =.032 v 3 (C) = max( , ) = o 1 o 2 o 3
21 Recall: An Example of the Viterbi Algorithmn v 1 (H) = 0.32 v 2 (H) = max(.32.12,.02.06) =.0384 v 3 (H) = max( , ) = S P(H S)P(3 H) P(C S)P(3 C) P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) P(H H)P(3 H) P(C H)P(3 C) P(H C)P(3 H) P(C C)P(3 C) C C C /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) =.032 v 3 (C) = max( , ) = o 1 o 2 o 3
22 Recall: An Example of the Viterbi Algorithmn S P(H S)P(3 H) P(C S)P(3 C) v 1 (H) = 0.32 P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) v 2 (H) = max(.32.12,.02.06) =.0384 P(H H)P(3 H) P(C H)P(3 C) P(H C)P(3 H) P(C C)P(3 C) v 3 (H) = max( , ) = C C C P( /S H) 0.2 v f ( /S ) = max( , ) = P( /S C) 0.2 /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) =.032 v 3 (C) = max( , ) = o 1 o 2 o 3
23 Recall: An Example of the Viterbi Algorithmn S P(H S)P(3 H) P(C S)P(3 C) v 1 (H) = 0.32 P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) v 2 (H) = max(.32.12,.02.06) =.0384 P(H H)P(3 H) P(C H)P(3 C) P(H C)P(3 H) P(C C)P(3 C) v 3 (H) = max( , ) = C C C P( /S H) 0.2 v f ( /S ) = max( , ) = P( /S C) 0.2 /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) = o 1 o 2 o 3 H v 3 (C) = max( , ) =.0016
24 Recall: An Example of the Viterbi Algorithmn S P(H S)P(3 H) P(C S)P(3 C) v 1 (H) = 0.32 P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) v 2 (H) = max(.32.12,.02.06) =.0384 P(H H)P(3 H) P(C H)P(3 C) P(H C)P(3 H) P(C C)P(3 C) v 3 (H) = max( , ) = C C C P( /S H) 0.2 v f ( /S ) = max( , ) = P( /S C) 0.2 /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) =.032 v 3 (C) = max( , ) = o 1 o 2 o 3 H H
25 Recall: An Example of the Viterbi Algorithmn S P(H S)P(3 H) P(C S)P(3 C) v 1 (H) = 0.32 P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) v 2 (H) = max(.32.12,.02.06) =.0384 P(H H)P(3 H) P(C H)P(3 C) P(H C)P(3 H) P(C C)P(3 C) v 3 (H) = max( , ) = C C C P( /S H) 0.2 v f ( /S ) = max( , ) = P( /S C) 0.2 /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) = o 1 o 2 o 3 H H H v 3 (C) = max( , ) =.0016
26 Recall: An Example of the Viterbi Algorithmn S P(H S)P(3 H) P(C S)P(3 C) v 1 (H) = 0.32 P(H H)P(1 H) H H H P(C H)P(1 C) P(H C)P(1 H) P(C C)P(1 C) v 2 (H) = max(.32.12,.02.06) =.0384 P(H H)P(3 H) P(C H)P(3 C) P(H C)P(3 H) P(C C)P(3 C) v 3 (H) = max( , ) = C C C P( /S H) 0.2 v f ( /S ) = max( , ) = P( /S C) 0.2 /S v 1 (C) = 0.02 v 2 (C) = max(.32.1,.02.25) = o 1 o 2 o 3 H H H v 3 (C) = max( , ) =.0016
27 Recall: Using HMMs The HMM models the process of generating the labelled sequence. We can use this model for a number of tasks: P(S, O) given S and O P(O) given O S that maximizes P(S O) given O P(s x O) given O We learn model parameters from a set of observations.
28 Moving Onwards Determining which string is most likely: How to recognize speech vs. How to wreck a nice beach which tag sequence is most likely for flies like flowers: NNS VB NNS vs. VBZ P NNS which syntactic structure is most likely: S S NP VP NP VP I VBD ate N sushi NP PP with tuna I VBD ate NP N sushi PP with tuna
29 Moving Onwards Determining which string is most likely: How to recognize speech vs. How to wreck a nice beach which tag sequence is most likely for flies like flowers: NNS VB NNS vs. VBZ P NNS which syntactic structure is most likely: S S NP VP NP VP I VBD ate N sushi NP PP with tuna I VBD ate NP N sushi PP with tuna
30 Moving Onwards Determining which string is most likely: How to recognize speech vs. How to wreck a nice beach which tag sequence is most likely for flies like flowers: NNS VB NNS vs. VBZ P NNS which syntactic structure is most likely: S S NP VP NP VP I VBD ate N sushi NP PP with tuna I VBD ate NP N sushi PP with tuna
31 Moving Onwards Determining which string is most likely: How to recognize speech vs. How to wreck a nice beach which tag sequence is most likely for flies like flowers: NNS VB NNS vs. VBZ P NNS which syntactic structure is most likely: S S NP VP NP VP I VBD ate N sushi NP PP with tuna I VBD ate NP N sushi PP with tuna
32 From Linear Order to Hierarchical Structure The models we have looked at so far: n-gram models (Markov chains). Purely linear (sequential) and surface-oriented. sequence labeling: HMMs. Adds one layer of abstraction: PoS as hidden variables. Still only sequential in nature.
33 From Linear Order to Hierarchical Structure The models we have looked at so far: n-gram models (Markov chains). Purely linear (sequential) and surface-oriented. sequence labeling: HMMs. Adds one layer of abstraction: PoS as hidden variables. Still only sequential in nature. Formal grammar adds hierarchical structure. In NLP, being a sub-discipline of AI, we want our programs to understand natural language (on some level). Finding the grammatical structure of sentences is an important step towards understanding. Shift focus from sequences to grammatical structures.
34 Why We Need Structure (1/3) Constituency Words tends to lump together into groups that behave like single units: we call them constituents. Constituency tests give evidence for constituent structure: interchangeable in similar syntactic environments. can be co-ordinated can be moved within a sentence as a unit
35 Why We Need Structure (1/3) Constituency Words tends to lump together into groups that behave like single units: we call them constituents. Constituency tests give evidence for constituent structure: interchangeable in similar syntactic environments. can be co-ordinated can be moved within a sentence as a unit (4) Kim read [a very interesting book about grammar] NP. Kim read [it] NP. (5) Kim [read a book] VP, [gave it to Sandy] VP, and [left] VP. (6) [Read the book] VP I really meant to this week. Examples from Linguistic Fundamentals for NLP: 100 Essentials from Morphology and Syntax. Bender (2013)
36 Why We Need Structure (2/3) Constituency Constituents are theory-dependent, and are not absolute or language-independent.
37 Why We Need Structure (2/3) Constituency Constituents are theory-dependent, and are not absolute or language-independent. A constituent usually has a head element, and is often named according to the type of its head: A noun phrase (NP) has a nominal (noun-type) head: (9) [ a very interesting book about grammar ] NP A verb phrase ) has a verbal head: (10) [ gives books to students ] VP
38 Why We Need Structure (3/3) Grammatical functions Terms such as subject and object describe the grammatical function of a constituent in a sentence.
39 Why We Need Structure (3/3) Grammatical functions Terms such as subject and object describe the grammatical function of a constituent in a sentence. Agreement establishes a symmetric relationship between grammatical features. The decision of the Nobel committee members surprises most of us. Why would a purely linear model have problems predicting this phenomenon?
40 Why We Need Structure (3/3) Grammatical functions Terms such as subject and object describe the grammatical function of a constituent in a sentence. Agreement establishes a symmetric relationship between grammatical features. The decision of the Nobel committee members surprises most of us. Why would a purely linear model have problems predicting this phenomenon? Verb agreement reflects the grammatical structure of the sentence, not just the sequential order of words.
41 Grammars: A Tool to Aid Understanding Formal grammars describe a language, giving us a way to: judge or predict well-formedness Kim was happy because passed the exam. Kim was happy because final grade was an A.
42 Grammars: A Tool to Aid Understanding Formal grammars describe a language, giving us a way to: judge or predict well-formedness Kim was happy because passed the exam. Kim was happy because final grade was an A. make explicit structural ambiguities Have her report on my desk by Friday! I like to eat sushi with { chopsticks tuna }.
43 Grammars: A Tool to Aid Understanding Formal grammars describe a language, giving us a way to: judge or predict well-formedness Kim was happy because passed the exam. Kim was happy because final grade was an A. make explicit structural ambiguities Have her report on my desk by Friday! I like to eat sushi with { chopsticks tuna }. derive abstract representations of meaning Kim gave Sandy a book. Kim gave a book to Sandy. Sandy was given a book by Kim.
44 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en
45 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en
46 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en
47 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
48 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
49 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
50 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
51 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
52 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
53 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
54 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
55 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
56 A Grossly Simplified Example The Grammar of Spanish S NP VP VP V NP VP VP PP PP P NP NP nieve NP Juan NP Oslo V amó P en S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
57 A Grossly Simplified Example The Grammar of Spanish S NP VP { VP ( NP ) } VP V NP { V ( NP ) } VP VP PP { PP ( VP ) } PP P NP { P ( NP ) } NP nieve { snow } NP Juan { John } NP Oslo { Oslo } V amó { λbλa adore ( a, b ) } P en { λdλc in ( c, d ) } S NP VP Juan VP PP V NP P NP amó nieve en Oslo Juan amó nieve en Oslo
58 Meaning Composition (Still Very Simplified) S: {in ( adore ( John, snow ), Oslo )} NP: {John} VP: {λa in ( adore ( a, snow ), Oslo )} Juan VP: {λa adore ( a, snow )} PP:{λc in ( c, Oslo )} V:{λbλa adore ( a, b )} NP:{snow} P:{λdλc in ( c, d )} NP:{Oslo} amó nieve en Oslo VP V NP { V ( NP ) }
59 Another Interpretation S: {adore (John, in ( snow, Oslo )} NP: {John} VP: {λa adore (a, in ( snow, Oslo )} Juan V:{λbλa adore ( a, b )} NP:{in ( snow, Oslo )} amó NP:{snow} PP:{λc in ( c, Oslo )} nieve P:{λdλc in ( c, d )} NP:{Oslo} en Oslo NP NP PP { PP ( NP ) }
60 Context Free Grammars (CFGs) Formal system for modeling constituent structure. Defined in terms of a lexicon and a set of rules
61 Context Free Grammars (CFGs) Formal system for modeling constituent structure. Defined in terms of a lexicon and a set of rules Formal models of language in a broad sense natural languages, programming languages, communication protocols,...
62 Context Free Grammars (CFGs) Formal system for modeling constituent structure. Defined in terms of a lexicon and a set of rules Formal models of language in a broad sense natural languages, programming languages, communication protocols,... Can be expressed in the meta-syntax of the Backus-Naur Form (BNF) formalism. When looking up concepts and syntax in the Common Lisp HyperSpec, you have been reading (extended) BNF.
63 Context Free Grammars (CFGs) Formal system for modeling constituent structure. Defined in terms of a lexicon and a set of rules Formal models of language in a broad sense natural languages, programming languages, communication protocols,... Can be expressed in the meta-syntax of the Backus-Naur Form (BNF) formalism. When looking up concepts and syntax in the Common Lisp HyperSpec, you have been reading (extended) BNF. Powerful enough to express sophisticated relations among words, yet in a computationally tractable way.
64 CFGs (Formally, this Time) Formally, a CFG is a quadruple: G = C, Σ, P, S
65 CFGs (Formally, this Time) Formally, a CFG is a quadruple: G = C, Σ, P, S C is the set of categories (aka non-terminals), {S, NP, VP, V}
66 CFGs (Formally, this Time) Formally, a CFG is a quadruple: G = C, Σ, P, S C is the set of categories (aka non-terminals), {S, NP, VP, V} Σ is the vocabulary (aka terminals), {Kim, snow, adores, in}
67 CFGs (Formally, this Time) Formally, a CFG is a quadruple: G = C, Σ, P, S C is the set of categories (aka non-terminals), {S, NP, VP, V} Σ is the vocabulary (aka terminals), {Kim, snow, adores, in} P is a set of category rewrite rules (aka productions) S NP VP VP V NP NP Kim NP snow V adores
68 CFGs (Formally, this Time) Formally, a CFG is a quadruple: G = C, Σ, P, S C is the set of categories (aka non-terminals), {S, NP, VP, V} Σ is the vocabulary (aka terminals), {Kim, snow, adores, in} P is a set of category rewrite rules (aka productions) S NP VP VP V NP NP Kim NP snow V adores S C is the start symbol, a filter on complete results;
69 CFGs (Formally, this Time) Formally, a CFG is a quadruple: G = C, Σ, P, S C is the set of categories (aka non-terminals), {S, NP, VP, V} Σ is the vocabulary (aka terminals), {Kim, snow, adores, in} P is a set of category rewrite rules (aka productions) S NP VP VP V NP NP Kim NP snow V adores S C is the start symbol, a filter on complete results; for each rule α β 1, β 2,..., β n P: α C and β i C Σ
70 Generative Grammar Top-down view of generative grammars: For a grammar G, the language L G is defined as the set of strings that can be derived from S. To derive w n from S, we use the rules in P to recursively 1 rewrite S into the sequence w n 1 where each w i Σ
71 Generative Grammar Top-down view of generative grammars: For a grammar G, the language L G is defined as the set of strings that can be derived from S. To derive w n from S, we use the rules in P to recursively 1 rewrite S into the sequence w n 1 where each w i Σ The grammar is seen as generating strings. Grammatical strings are defined as strings that can be generated by the grammar.
72 Generative Grammar Top-down view of generative grammars: For a grammar G, the language L G is defined as the set of strings that can be derived from S. To derive w n from S, we use the rules in P to recursively 1 rewrite S into the sequence w n 1 where each w i Σ The grammar is seen as generating strings. Grammatical strings are defined as strings that can be generated by the grammar. The context-freeness of CFGs refers to the fact that we rewrite non-terminals without regard to the overall context in which they occur.
73 Treebanks Generally A treebank is a corpus paired with gold-standard (syntactico-semantic) analyses Can be created by manual annotation or selection among outputs from automated processing (plus correction).
74 Treebanks Generally A treebank is a corpus paired with gold-standard (syntactico-semantic) analyses Can be created by manual annotation or selection among outputs from automated processing (plus correction). Penn Treebank (Marcus et al., 1993) About one million tokens of Wall Street Journal text Hand-corrected PoS annotation using 45 word classes Manual annotation with (somewhat) coarse constituent structure
75 One Example from the Penn Treebank S [WSJ 2350] advp, np-sbj-1 vp. rb, np nn vbz vp. Still nnp pos move is vbg vbn np advp-mnr Time s being received -none- rb Still, Time s move is being received well. *-1 well
76 One Example from the Penn Treebank S [WSJ 2350] advp, np-sbj-1 vp. rb, np nn vbz vp. Still nnp pos move is vbg vbn np advp-mnr Time s being received -none- rb Still, Time s move is being received well. *-1 well
77 One Example from the Penn Treebank S [WSJ 2350] advp, np-sbj-1 vp. rb, np nn vbz vp. Still nnp pos move is vbg vbn np advp-mnr Time s being received -none- rb Still, Time s move is being received well. *-1 well
78 Elimination of Traces and Functions S [WSJ 2350] advp, np vp. rb, np nn vbz vp. Still nnp pos move is vbg vbn advp Time s being received rb Still, Time s move is being received well. well
79 Probabilitic Context-Free Grammars We are interested, not just in which trees apply to a sentence, but also to which tree is most likely.
80 Probabilitic Context-Free Grammars We are interested, not just in which trees apply to a sentence, but also to which tree is most likely. Probabilistic context-free grammars (PCFGs) augment CFGs by adding probabilities to each production, e.g. S NP VP 0.6 S NP VP PP 0.4 These are conditional probabilities the probability of the right hand side (RHS) given the left hand side (LHS) P(S NP VP) = P(NP VP S)
81 Probabilitic Context-Free Grammars We are interested, not just in which trees apply to a sentence, but also to which tree is most likely. Probabilistic context-free grammars (PCFGs) augment CFGs by adding probabilities to each production, e.g. S NP VP 0.6 S NP VP PP 0.4 These are conditional probabilities the probability of the right hand side (RHS) given the left hand side (LHS) P(S NP VP) = P(NP VP S) We can learn these probabilities from a treebank, again using Maximum Likelihood Estimation.
82 Estimating PCFGs (1/3) S [WSJ 2350] advp, np vp. rb, np nn vbz vp. Still nnp pos move is vbg vbn advp Time s being received rb Still, Time s move is being received well. well
83 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. "."))
84 Estimating PCFGs (2/3) RB Still 1 (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. "."))
85 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1
86 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1,, 1
87 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1,, 1 NNP Time 1
88 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1,, 1 NNP Time 1 POS s 1
89 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1,, 1 NNP Time 1 POS s 1 NP NNP POS 1
90 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1
91 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1
92 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1 VBZ is 1
93 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1 VBZ is 1 VBG being 1
94 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1 VBZ is 1 VBG being 1 VBN received 1
95 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 1,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1 VBZ is 1 VBG being 1 VBN received 1 RB well 1
96 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 2,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1 VBZ is 1 VBG being 1 VBN received 1 RB well 1
97 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 2,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1 VBZ is 1 VBG being 1 VBN received 1 RB well 1 VP VBN ADVP 1
98 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 2,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1 VBZ is 1 VBG being 1 VBN received 1 RB well 1 VP VBN ADVP 1 VP VBG VP 1
99 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 2,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1 VBZ is 1 VBG being 1 VBN received 1 RB well 1 VP VBN ADVP 1 VP VBG VP 1 \.. 1
100 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 2,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1 VBZ is 1 VBG being 1 VBN received 1 RB well 1 VP VBN ADVP 1 VP VBG VP 1 \.. 1 S ADVP, NP VP \. 1
101 Estimating PCFGs (2/3) (S (ADVP (RB "Still")) (, ",") (NP (NP (NNP "Time") (POS " s")) (NN "move")) (VBZ "is") (VBG "being") (VBN "received") (ADVP (RB "well"))))) (\. ".")) RB Still 1 ADVP RB 2,, 1 NNP Time 1 POS s 1 NP NNP POS 1 NN move 1 NP NP NN 1 VBZ is 1 VBG being 1 VBN received 1 RB well 1 VP VBN ADVP 1 VP VBG VP 1 \.. 1 S ADVP, NP VP \. 1 START S 1
102 Estimating PCFGs (3/3) Once we have counts of all the rules, we turn them into probabilities.
103 Estimating PCFGs (3/3) Once we have counts of all the rules, we turn them into probabilities. S ADVP, NP VP \. 50 S NP VP \. 400 S NP VP PP \. 350 S VP! 100 S NP VP S \. 200 S NP VP 50
104 Estimating PCFGs (3/3) Once we have counts of all the rules, we turn them into probabilities. S ADVP, NP VP \. 50 S NP VP \. 400 S NP VP PP \. 350 S VP! 100 S NP VP S \. 200 S NP VP 50 P(S ADVP, NP VP \.) C(S ADVP, NP VP \.) C(S)
105 Estimating PCFGs (3/3) Once we have counts of all the rules, we turn them into probabilities. S ADVP, NP VP \. 50 S NP VP \. 400 S NP VP PP \. 350 S VP! 100 S NP VP S \. 200 S NP VP 50 P(S ADVP, NP VP \.) = C(S ADVP, NP VP \.) C(S) =
Outline of today s lecture
Outline of today s lecture Generative grammar Simple context free grammars Probabilistic CFGs Formalism power requirements Parsing Modelling syntactic structure of phrases and sentences. Why is it useful?
More informationPOS Tagsets and POS Tagging. Definition. Tokenization. Tagset Design. Automatic POS Tagging Bigram tagging. Maximum Likelihood Estimation 1 / 23
POS Def. Part of Speech POS POS L645 POS = Assigning word class information to words Dept. of Linguistics, Indiana University Fall 2009 ex: the man bought a book determiner noun verb determiner noun 1
More informationTagging with Hidden Markov Models
Tagging with Hidden Markov Models Michael Collins 1 Tagging Problems In many NLP problems, we would like to model pairs of sequences. Part-of-speech (POS) tagging is perhaps the earliest, and most famous,
More informationAccelerating and Evaluation of Syntactic Parsing in Natural Language Question Answering Systems
Accelerating and Evaluation of Syntactic Parsing in Natural Language Question Answering Systems cation systems. For example, NLP could be used in Question Answering (QA) systems to understand users natural
More informationIntroduction. BM1 Advanced Natural Language Processing. Alexander Koller. 17 October 2014
Introduction! BM1 Advanced Natural Language Processing Alexander Koller! 17 October 2014 Outline What is computational linguistics? Topics of this course Organizational issues Siri Text prediction Facebook
More informationSymbiosis of Evolutionary Techniques and Statistical Natural Language Processing
1 Symbiosis of Evolutionary Techniques and Statistical Natural Language Processing Lourdes Araujo Dpto. Sistemas Informáticos y Programación, Univ. Complutense, Madrid 28040, SPAIN (email: lurdes@sip.ucm.es)
More informationINF5820 Natural Language Processing - NLP. H2009 Jan Tore Lønning jtl@ifi.uio.no
INF5820 Natural Language Processing - NLP H2009 Jan Tore Lønning jtl@ifi.uio.no Semantic Role Labeling INF5830 Lecture 13 Nov 4, 2009 Today Some words about semantics Thematic/semantic roles PropBank &
More informationSyntax: Phrases. 1. The phrase
Syntax: Phrases Sentences can be divided into phrases. A phrase is a group of words forming a unit and united around a head, the most important part of the phrase. The head can be a noun NP, a verb VP,
More informationWhy language is hard. And what Linguistics has to say about it. Natalia Silveira Participation code: eagles
Why language is hard And what Linguistics has to say about it Natalia Silveira Participation code: eagles Christopher Natalia Silveira Manning Language processing is so easy for humans that it is like
More informationChunk Parsing. Steven Bird Ewan Klein Edward Loper. University of Melbourne, AUSTRALIA. University of Edinburgh, UK. University of Pennsylvania, USA
Chunk Parsing Steven Bird Ewan Klein Edward Loper University of Melbourne, AUSTRALIA University of Edinburgh, UK University of Pennsylvania, USA March 1, 2012 chunk parsing: efficient and robust approach
More informationParsing Technology and its role in Legacy Modernization. A Metaware White Paper
Parsing Technology and its role in Legacy Modernization A Metaware White Paper 1 INTRODUCTION In the two last decades there has been an explosion of interest in software tools that can automate key tasks
More informationGrammars and introduction to machine learning. Computers Playing Jeopardy! Course Stony Brook University
Grammars and introduction to machine learning Computers Playing Jeopardy! Course Stony Brook University Last class: grammars and parsing in Prolog Noun -> roller Verb thrills VP Verb NP S NP VP NP S VP
More informationA Mixed Trigrams Approach for Context Sensitive Spell Checking
A Mixed Trigrams Approach for Context Sensitive Spell Checking Davide Fossati and Barbara Di Eugenio Department of Computer Science University of Illinois at Chicago Chicago, IL, USA dfossa1@uic.edu, bdieugen@cs.uic.edu
More informationNATURAL LANGUAGE QUERY PROCESSING USING PROBABILISTIC CONTEXT FREE GRAMMAR
NATURAL LANGUAGE QUERY PROCESSING USING PROBABILISTIC CONTEXT FREE GRAMMAR Arati K. Deshpande 1 and Prakash. R. Devale 2 1 Student and 2 Professor & Head, Department of Information Technology, Bharati
More informationNLP Programming Tutorial 5 - Part of Speech Tagging with Hidden Markov Models
NLP Programming Tutorial 5 - Part of Speech Tagging with Hidden Markov Models Graham Neubig Nara Institute of Science and Technology (NAIST) 1 Part of Speech (POS) Tagging Given a sentence X, predict its
More informationAutomatic Detection and Correction of Errors in Dependency Treebanks
Automatic Detection and Correction of Errors in Dependency Treebanks Alexander Volokh DFKI Stuhlsatzenhausweg 3 66123 Saarbrücken, Germany alexander.volokh@dfki.de Günter Neumann DFKI Stuhlsatzenhausweg
More information31 Case Studies: Java Natural Language Tools Available on the Web
31 Case Studies: Java Natural Language Tools Available on the Web Chapter Objectives Chapter Contents This chapter provides a number of sources for open source and free atural language understanding software
More informationStatistical Machine Translation: IBM Models 1 and 2
Statistical Machine Translation: IBM Models 1 and 2 Michael Collins 1 Introduction The next few lectures of the course will be focused on machine translation, and in particular on statistical machine translation
More informationText Chunking based on a Generalization of Winnow
Journal of Machine Learning Research 2 (2002) 615-637 Submitted 9/01; Published 3/02 Text Chunking based on a Generalization of Winnow Tong Zhang Fred Damerau David Johnson T.J. Watson Research Center
More informationCINTIL-PropBank. CINTIL-PropBank Sub-corpus id Sentences Tokens Domain Sentences for regression atsts 779 5,654 Test
CINTIL-PropBank I. Basic Information 1.1. Corpus information The CINTIL-PropBank (Branco et al., 2012) is a set of sentences annotated with their constituency structure and semantic role tags, composed
More informationPP-Attachment. Chunk/Shallow Parsing. Chunk Parsing. PP-Attachment. Recall the PP-Attachment Problem (demonstrated with XLE):
PP-Attachment Recall the PP-Attachment Problem (demonstrated with XLE): Chunk/Shallow Parsing The girl saw the monkey with the telescope. 2 readings The ambiguity increases exponentially with each PP.
More informationTesting Data-Driven Learning Algorithms for PoS Tagging of Icelandic
Testing Data-Driven Learning Algorithms for PoS Tagging of Icelandic by Sigrún Helgadóttir Abstract This paper gives the results of an experiment concerned with training three different taggers on tagged
More informationHow the Computer Translates. Svetlana Sokolova President and CEO of PROMT, PhD.
Svetlana Sokolova President and CEO of PROMT, PhD. How the Computer Translates Machine translation is a special field of computer application where almost everyone believes that he/she is a specialist.
More informationAutomatic Text Analysis Using Drupal
Automatic Text Analysis Using Drupal By Herman Chai Computer Engineering California Polytechnic State University, San Luis Obispo Advised by Dr. Foaad Khosmood June 14, 2013 Abstract Natural language processing
More informationLearning Translation Rules from Bilingual English Filipino Corpus
Proceedings of PACLIC 19, the 19 th Asia-Pacific Conference on Language, Information and Computation. Learning Translation s from Bilingual English Filipino Corpus Michelle Wendy Tan, Raymond Joseph Ang,
More informationApplying Co-Training Methods to Statistical Parsing. Anoop Sarkar http://www.cis.upenn.edu/ anoop/ anoop@linc.cis.upenn.edu
Applying Co-Training Methods to Statistical Parsing Anoop Sarkar http://www.cis.upenn.edu/ anoop/ anoop@linc.cis.upenn.edu 1 Statistical Parsing: the company s clinical trials of both its animal and human-based
More informationBasic Parsing Algorithms Chart Parsing
Basic Parsing Algorithms Chart Parsing Seminar Recent Advances in Parsing Technology WS 2011/2012 Anna Schmidt Talk Outline Chart Parsing Basics Chart Parsing Algorithms Earley Algorithm CKY Algorithm
More informationDEPENDENCY PARSING JOAKIM NIVRE
DEPENDENCY PARSING JOAKIM NIVRE Contents 1. Dependency Trees 1 2. Arc-Factored Models 3 3. Online Learning 3 4. Eisner s Algorithm 4 5. Spanning Tree Parsing 6 References 7 A dependency parser analyzes
More informationCourse: Model, Learning, and Inference: Lecture 5
Course: Model, Learning, and Inference: Lecture 5 Alan Yuille Department of Statistics, UCLA Los Angeles, CA 90095 yuille@stat.ucla.edu Abstract Probability distributions on structured representation.
More informationHidden Markov Models Chapter 15
Hidden Markov Models Chapter 15 Mausam (Slides based on Dan Klein, Luke Zettlemoyer, Alex Simma, Erik Sudderth, David Fernandez-Baca, Drena Dobbs, Serafim Batzoglou, William Cohen, Andrew McCallum, Dan
More informationShallow Parsing with Apache UIMA
Shallow Parsing with Apache UIMA Graham Wilcock University of Helsinki Finland graham.wilcock@helsinki.fi Abstract Apache UIMA (Unstructured Information Management Architecture) is a framework for linguistic
More informationWhat s in a Lexicon. The Lexicon. Lexicon vs. Dictionary. What kind of Information should a Lexicon contain?
What s in a Lexicon What kind of Information should a Lexicon contain? The Lexicon Miriam Butt November 2002 Semantic: information about lexical meaning and relations (thematic roles, selectional restrictions,
More informationNATURAL LANGUAGE QUERY PROCESSING USING SEMANTIC GRAMMAR
NATURAL LANGUAGE QUERY PROCESSING USING SEMANTIC GRAMMAR 1 Gauri Rao, 2 Chanchal Agarwal, 3 Snehal Chaudhry, 4 Nikita Kulkarni,, 5 Dr. S.H. Patil 1 Lecturer department o f Computer Engineering BVUCOE,
More informationNoam Chomsky: Aspects of the Theory of Syntax notes
Noam Chomsky: Aspects of the Theory of Syntax notes Julia Krysztofiak May 16, 2006 1 Methodological preliminaries 1.1 Generative grammars as theories of linguistic competence The study is concerned with
More informationThe Specific Text Analysis Tasks at the Beginning of MDA Life Cycle
SCIENTIFIC PAPERS, UNIVERSITY OF LATVIA, 2010. Vol. 757 COMPUTER SCIENCE AND INFORMATION TECHNOLOGIES 11 22 P. The Specific Text Analysis Tasks at the Beginning of MDA Life Cycle Armands Šlihte Faculty
More informationEffective Self-Training for Parsing
Effective Self-Training for Parsing David McClosky dmcc@cs.brown.edu Brown Laboratory for Linguistic Information Processing (BLLIP) Joint work with Eugene Charniak and Mark Johnson David McClosky - dmcc@cs.brown.edu
More informationA Knowledge-based System for Translating FOL Formulas into NL Sentences
A Knowledge-based System for Translating FOL Formulas into NL Sentences Aikaterini Mpagouli, Ioannis Hatzilygeroudis University of Patras, School of Engineering Department of Computer Engineering & Informatics,
More informationOpen Domain Information Extraction. Günter Neumann, DFKI, 2012
Open Domain Information Extraction Günter Neumann, DFKI, 2012 Improving TextRunner Wu and Weld (2010) Open Information Extraction using Wikipedia, ACL 2010 Fader et al. (2011) Identifying Relations for
More informationEnglish Grammar Checker
International l Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Issue-3 E-ISSN: 2347-2693 English Grammar Checker Pratik Ghosalkar 1*, Sarvesh Malagi 2, Vatsal Nagda 3,
More informationShallow Parsing with PoS Taggers and Linguistic Features
Journal of Machine Learning Research 2 (2002) 639 668 Submitted 9/01; Published 3/02 Shallow Parsing with PoS Taggers and Linguistic Features Beáta Megyesi Centre for Speech Technology (CTT) Department
More informationCS 6740 / INFO 6300. Ad-hoc IR. Graduate-level introduction to technologies for the computational treatment of information in humanlanguage
CS 6740 / INFO 6300 Advanced d Language Technologies Graduate-level introduction to technologies for the computational treatment of information in humanlanguage form, covering natural-language processing
More informationArtificial Intelligence Exam DT2001 / DT2006 Ordinarie tentamen
Artificial Intelligence Exam DT2001 / DT2006 Ordinarie tentamen Date: 2010-01-11 Time: 08:15-11:15 Teacher: Mathias Broxvall Phone: 301438 Aids: Calculator and/or a Swedish-English dictionary Points: The
More informationAutomatic Knowledge Base Construction Systems. Dr. Daisy Zhe Wang CISE Department University of Florida September 3th 2014
Automatic Knowledge Base Construction Systems Dr. Daisy Zhe Wang CISE Department University of Florida September 3th 2014 1 Text Contains Knowledge 2 Text Contains Automatically Extractable Knowledge 3
More informationSpecial Topics in Computer Science
Special Topics in Computer Science NLP in a Nutshell CS492B Spring Semester 2009 Jong C. Park Computer Science Department Korea Advanced Institute of Science and Technology INTRODUCTION Jong C. Park, CS
More informationImproving Data Driven Part-of-Speech Tagging by Morphologic Knowledge Induction
Improving Data Driven Part-of-Speech Tagging by Morphologic Knowledge Induction Uwe D. Reichel Department of Phonetics and Speech Communication University of Munich reichelu@phonetik.uni-muenchen.de Abstract
More informationLing 201 Syntax 1. Jirka Hana April 10, 2006
Overview of topics What is Syntax? Word Classes What to remember and understand: Ling 201 Syntax 1 Jirka Hana April 10, 2006 Syntax, difference between syntax and semantics, open/closed class words, all
More informationMachine Learning for natural language processing
Machine Learning for natural language processing Introduction Laura Kallmeyer Heinrich-Heine-Universität Düsseldorf Summer 2016 1 / 13 Introduction Goal of machine learning: Automatically learn how to
More informationConditional Random Fields: An Introduction
Conditional Random Fields: An Introduction Hanna M. Wallach February 24, 2004 1 Labeling Sequential Data The task of assigning label sequences to a set of observation sequences arises in many fields, including
More informationNgram Search Engine with Patterns Combining Token, POS, Chunk and NE Information
Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information Satoshi Sekine Computer Science Department New York University sekine@cs.nyu.edu Kapil Dalwani Computer Science Department
More informationModule Catalogue for the Bachelor Program in Computational Linguistics at the University of Heidelberg
Module Catalogue for the Bachelor Program in Computational Linguistics at the University of Heidelberg March 1, 2007 The catalogue is organized into sections of (1) obligatory modules ( Basismodule ) that
More informationNatural Language Processing
Natural Language Processing 2 Open NLP (http://opennlp.apache.org/) Java library for processing natural language text Based on Machine Learning tools maximum entropy, perceptron Includes pre-built models
More informationSyntactic Theory. Background and Transformational Grammar. Dr. Dan Flickinger & PD Dr. Valia Kordoni
Syntactic Theory Background and Transformational Grammar Dr. Dan Flickinger & PD Dr. Valia Kordoni Department of Computational Linguistics Saarland University October 28, 2011 Early work on grammar There
More informationHow To Understand A Sentence In A Syntactic Analysis
AN AUGMENTED STATE TRANSITION NETWORK ANALYSIS PROCEDURE Daniel G. Bobrow Bolt, Beranek and Newman, Inc. Cambridge, Massachusetts Bruce Eraser Language Research Foundation Cambridge, Massachusetts Summary
More informationStatistical Machine Translation
Statistical Machine Translation Some of the content of this lecture is taken from previous lectures and presentations given by Philipp Koehn and Andy Way. Dr. Jennifer Foster National Centre for Language
More informationSearch and Data Mining: Techniques. Text Mining Anya Yarygina Boris Novikov
Search and Data Mining: Techniques Text Mining Anya Yarygina Boris Novikov Introduction Generally used to denote any system that analyzes large quantities of natural language text and detects lexical or
More informationMotivation. Korpus-Abfrage: Werkzeuge und Sprachen. Overview. Languages of Corpus Query. SARA Query Possibilities 1
Korpus-Abfrage: Werkzeuge und Sprachen Gastreferat zur Vorlesung Korpuslinguistik mit und für Computerlinguistik Charlotte Merz 3. Dezember 2002 Motivation Lizentiatsarbeit: A Corpus Query Tool for Automatically
More informationUsing classes has the potential of reducing the problem of sparseness of data by allowing generalizations
POS Tags and Decision Trees for Language Modeling Peter A. Heeman Department of Computer Science and Engineering Oregon Graduate Institute PO Box 91000, Portland OR 97291 heeman@cse.ogi.edu Abstract Language
More information2 Related Work and Overview of Classification Techniques
Document Classification for Newspaper Articles Dennis Ramdass & Shreyes Seshasai 6.863 Final Project Spring 2009 May 18, 2009 1 Introduction In many real-world scenarios, the ability to automatically classify
More informationLog-Linear Models. Michael Collins
Log-Linear Models Michael Collins 1 Introduction This note describes log-linear models, which are very widely used in natural language processing. A key advantage of log-linear models is their flexibility:
More informationSo today we shall continue our discussion on the search engines and web crawlers. (Refer Slide Time: 01:02)
Internet Technology Prof. Indranil Sengupta Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur Lecture No #39 Search Engines and Web Crawler :: Part 2 So today we
More informationBILINGUAL TRANSLATION SYSTEM
BILINGUAL TRANSLATION SYSTEM (FOR ENGLISH AND TAMIL) Dr. S. Saraswathi Associate Professor M. Anusiya P. Kanivadhana S. Sathiya Abstract--- The project aims in developing Bilingual Translation System for
More informationSpeech Recognition on Cell Broadband Engine UCRL-PRES-223890
Speech Recognition on Cell Broadband Engine UCRL-PRES-223890 Yang Liu, Holger Jones, John Johnson, Sheila Vaidya (Lawrence Livermore National Laboratory) Michael Perrone, Borivoj Tydlitat, Ashwini Nanda
More informationHidden Markov Models in Bioinformatics. By Máthé Zoltán Kőrösi Zoltán 2006
Hidden Markov Models in Bioinformatics By Máthé Zoltán Kőrösi Zoltán 2006 Outline Markov Chain HMM (Hidden Markov Model) Hidden Markov Models in Bioinformatics Gene Finding Gene Finding Model Viterbi algorithm
More informationLog-Linear Models a.k.a. Logistic Regression, Maximum Entropy Models
Log-Linear Models a.k.a. Logistic Regression, Maximum Entropy Models Natural Language Processing CS 6120 Spring 2014 Northeastern University David Smith (some slides from Jason Eisner and Dan Klein) summary
More informationActivities. but I will require that groups present research papers
CS-498 Signals AI Themes Much of AI occurs at the signal level Processing data and making inferences rather than logical reasoning Areas such as vision, speech, NLP, robotics methods bleed into other areas
More informationINTEGRATED DEVELOPMENT ENVIRONMENTS FOR NATURAL LANGUAGE PROCESSING
INTEGRATED DEVELOPMENT ENVIRONMENTS FOR NATURAL LANGUAGE PROCESSING Text Analysis International, Inc. 1669-2 Hollenbeck Ave. #501 Sunnyvale, California 94087 October 2001 INTRODUCTION From the time of
More informationEfficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words
, pp.290-295 http://dx.doi.org/10.14257/astl.2015.111.55 Efficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words Irfan
More informationStructure of Presentation. The Role of Programming in Informatics Curricula. Concepts of Informatics 2. Concepts of Informatics 1
The Role of Programming in Informatics Curricula A. J. Cowling Department of Computer Science University of Sheffield Structure of Presentation Introduction The problem, and the key concepts. Dimensions
More informationSpecialty Answering Service. All rights reserved.
0 Contents 1 Introduction... 2 1.1 Types of Dialog Systems... 2 2 Dialog Systems in Contact Centers... 4 2.1 Automated Call Centers... 4 3 History... 3 4 Designing Interactive Dialogs with Structured Data...
More informationPOS Tagging 1. POS Tagging. Rule-based taggers Statistical taggers Hybrid approaches
POS Tagging 1 POS Tagging Rule-based taggers Statistical taggers Hybrid approaches POS Tagging 1 POS Tagging 2 Words taken isolatedly are ambiguous regarding its POS Yo bajo con el hombre bajo a PP AQ
More informationThe multilayer sentiment analysis model based on Random forest Wei Liu1, Jie Zhang2
2nd International Conference on Advances in Mechanical Engineering and Industrial Informatics (AMEII 2016) The multilayer sentiment analysis model based on Random forest Wei Liu1, Jie Zhang2 1 School of
More informationLanguage Model of Parsing and Decoding
Syntax-based Language Models for Statistical Machine Translation Eugene Charniak ½, Kevin Knight ¾ and Kenji Yamada ¾ ½ Department of Computer Science, Brown University ¾ Information Sciences Institute,
More informationA Chart Parsing implementation in Answer Set Programming
A Chart Parsing implementation in Answer Set Programming Ismael Sandoval Cervantes Ingenieria en Sistemas Computacionales ITESM, Campus Guadalajara elcoraz@gmail.com Rogelio Dávila Pérez Departamento de
More informationMulti-Engine Machine Translation by Recursive Sentence Decomposition
Multi-Engine Machine Translation by Recursive Sentence Decomposition Bart Mellebeek Karolina Owczarzak Josef van Genabith Andy Way National Centre for Language Technology School of Computing Dublin City
More informationChinese Open Relation Extraction for Knowledge Acquisition
Chinese Open Relation Extraction for Knowledge Acquisition Yuen-Hsien Tseng 1, Lung-Hao Lee 1,2, Shu-Yen Lin 1, Bo-Shun Liao 1, Mei-Jun Liu 1, Hsin-Hsi Chen 2, Oren Etzioni 3, Anthony Fader 4 1 Information
More informationBuilding a Question Classifier for a TREC-Style Question Answering System
Building a Question Classifier for a TREC-Style Question Answering System Richard May & Ari Steinberg Topic: Question Classification We define Question Classification (QC) here to be the task that, given
More informationDesign Patterns in Parsing
Abstract Axel T. Schreiner Department of Computer Science Rochester Institute of Technology 102 Lomb Memorial Drive Rochester NY 14623-5608 USA ats@cs.rit.edu Design Patterns in Parsing James E. Heliotis
More informationWe have recently developed a video database
Qibin Sun Infocomm Research A Natural Language-Based Interface for Querying a Video Database Onur Küçüktunç, Uğur Güdükbay, Özgür Ulusoy Bilkent University We have recently developed a video database system,
More informationIntroduction. Philipp Koehn. 28 January 2016
Introduction Philipp Koehn 28 January 2016 Administrativa 1 Class web site: http://www.mt-class.org/jhu/ Tuesdays and Thursdays, 1:30-2:45, Hodson 313 Instructor: Philipp Koehn (with help from Matt Post)
More informationAlgorithm & Flowchart & Pseudo code. Staff Incharge: S.Sasirekha
Algorithm & Flowchart & Pseudo code Staff Incharge: S.Sasirekha Computer Programming and Languages Computers work on a set of instructions called computer program, which clearly specify the ways to carry
More informationNatural Language Database Interface for the Community Based Monitoring System *
Natural Language Database Interface for the Community Based Monitoring System * Krissanne Kaye Garcia, Ma. Angelica Lumain, Jose Antonio Wong, Jhovee Gerard Yap, Charibeth Cheng De La Salle University
More informationAn NLP Curator (or: How I Learned to Stop Worrying and Love NLP Pipelines)
An NLP Curator (or: How I Learned to Stop Worrying and Love NLP Pipelines) James Clarke, Vivek Srikumar, Mark Sammons, Dan Roth Department of Computer Science, University of Illinois, Urbana-Champaign.
More informationOverview of MT techniques. Malek Boualem (FT)
Overview of MT techniques Malek Boualem (FT) This section presents an standard overview of general aspects related to machine translation with a description of different techniques: bilingual, transfer,
More informationNatural Language to Relational Query by Using Parsing Compiler
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 3, March 2015,
More informationContext Grammar and POS Tagging
Context Grammar and POS Tagging Shian-jung Dick Chen Don Loritz New Technology and Research New Technology and Research LexisNexis LexisNexis Ohio, 45342 Ohio, 45342 dick.chen@lexisnexis.com don.loritz@lexisnexis.com
More information209 THE STRUCTURE AND USE OF ENGLISH.
209 THE STRUCTURE AND USE OF ENGLISH. (3) A general survey of the history, structure, and use of the English language. Topics investigated include: the history of the English language; elements of the
More informationSemantic annotation of requirements for automatic UML class diagram generation
www.ijcsi.org 259 Semantic annotation of requirements for automatic UML class diagram generation Soumaya Amdouni 1, Wahiba Ben Abdessalem Karaa 2 and Sondes Bouabid 3 1 University of tunis High Institute
More informationUNKNOWN WORDS ANALYSIS IN POS TAGGING OF SINHALA LANGUAGE
UNKNOWN WORDS ANALYSIS IN POS TAGGING OF SINHALA LANGUAGE A.J.P.M.P. Jayaweera #1, N.G.J. Dias *2 # Virtusa Pvt. Ltd. No 752, Dr. Danister De Silva Mawatha, Colombo 09, Sri Lanka * Department of Statistics
More informationPhrase Structure Rules, Tree Rewriting, and other sources of Recursion Structure within the NP
Introduction to Transformational Grammar, LINGUIST 601 September 14, 2006 Phrase Structure Rules, Tree Rewriting, and other sources of Recursion Structure within the 1 Trees (1) a tree for the brown fox
More informationVisualizing Data Structures in Parsing-based Machine Translation. Jonathan Weese, Chris Callison-Burch
The Prague Bulletin of Mathematical Linguistics NUMBER 93 JANUARY 2010 127 136 Visualizing Data Structures in Parsing-based Machine Translation Jonathan Weese, Chris Callison-Burch Abstract As machine
More informationIntroduction to formal semantics -
Introduction to formal semantics - Introduction to formal semantics 1 / 25 structure Motivation - Philosophy paradox antinomy division in object und Meta language Semiotics syntax semantics Pragmatics
More informationTeaching Formal Methods for Computational Linguistics at Uppsala University
Teaching Formal Methods for Computational Linguistics at Uppsala University Roussanka Loukanova Computational Linguistics Dept. of Linguistics and Philology, Uppsala University P.O. Box 635, 751 26 Uppsala,
More informationExtraction of Legal Definitions from a Japanese Statutory Corpus Toward Construction of a Legal Term Ontology
Extraction of Legal Definitions from a Japanese Statutory Corpus Toward Construction of a Legal Term Ontology Makoto Nakamura, Yasuhiro Ogawa, Katsuhiko Toyama Japan Legal Information Institute, Graduate
More informationBing Liu. Web Data Mining. Exploring Hyperlinks, Contents, and Usage Data. With 177 Figures
Bing Liu Web Data Mining Exploring Hyperlinks, Contents, and Usage Data With 177 Figures 123 11 Opinion Mining In Chap. 9, we studied structured data extraction from Web pages. Such data are usually records
More informationParaphrasing controlled English texts
Paraphrasing controlled English texts Kaarel Kaljurand Institute of Computational Linguistics, University of Zurich kaljurand@gmail.com Abstract. We discuss paraphrasing controlled English texts, by defining
More informationWeb Document Clustering
Web Document Clustering Lab Project based on the MDL clustering suite http://www.cs.ccsu.edu/~markov/mdlclustering/ Zdravko Markov Computer Science Department Central Connecticut State University New Britain,
More informationPharmacovigilance, also referred to as drug safety surveillance, has been
SOCIAL MEDIA Identifying Adverse Drug Events from Patient Social Media A Case Study for Diabetes Xiao Liu, University of Arizona Hsinchun Chen, University of Arizona and Tsinghua University Social media
More informationOverview of the TACITUS Project
Overview of the TACITUS Project Jerry R. Hobbs Artificial Intelligence Center SRI International 1 Aims of the Project The specific aim of the TACITUS project is to develop interpretation processes for
More informationConstraints in Phrase Structure Grammar
Constraints in Phrase Structure Grammar Phrase Structure Grammar no movement, no transformations, context-free rules X/Y = X is a category which dominates a missing category Y Let G be the set of basic
More informationLanguage Modeling. Chapter 1. 1.1 Introduction
Chapter 1 Language Modeling (Course notes for NLP by Michael Collins, Columbia University) 1.1 Introduction In this chapter we will consider the the problem of constructing a language model from a set
More information