Statistical NLP Spring 2008. Machine Translation: Examples



Similar documents
Statistical Machine Translation Lecture 4. Beyond IBM Model 1 to Phrase-Based Models

Machine Translation. Agenda

TS3: an Improved Version of the Bilingual Concordancer TransSearch

Statistical Machine Translation: IBM Models 1 and 2

Chapter 5. Phrase-based models. Statistical Machine Translation

Convergence of Translation Memory and Statistical Machine Translation

An Iteratively-Trained Segmentation-Free Phrase Translation Model for Statistical Machine Translation

Chapter 6. Decoding. Statistical Machine Translation

IRIS - English-Irish Translation System

A Systematic Comparison of Various Statistical Alignment Models

Phrase-Based MT. Machine Translation Lecture 7. Instructor: Chris Callison-Burch TAs: Mitchell Stern, Justin Chiu. Website: mt-class.

The XMU Phrase-Based Statistical Machine Translation System for IWSLT 2006

Adaptive Development Data Selection for Log-linear Model in Statistical Machine Translation

Statistical Pattern-Based Machine Translation with Statistical French-English Machine Translation

Statistical Machine Translation of French and German into English Using IBM Model 2 Greedy Decoding

Statistical Machine Translation

Genetic Algorithm-based Multi-Word Automatic Language Translation

Factored Language Models for Statistical Machine Translation

Why Evaluation? Machine Translation. Evaluation. Evaluation Metrics. Ten Translations of a Chinese Sentence. How good is a given system?

SYSTRAN Chinese-English and English-Chinese Hybrid Machine Translation Systems for CWMT2011 SYSTRAN 混 合 策 略 汉 英 和 英 汉 机 器 翻 译 系 CWMT2011 技 术 报 告

Machine Translation. Why Evaluation? Evaluation. Ten Translations of a Chinese Sentence. Evaluation Metrics. But MT evaluation is a di cult problem!

Deciphering Foreign Language

A Joint Sequence Translation Model with Integrated Reordering

M LTO Multilingual On-Line Translation

tance alignment and time information to create confusion networks 1 from the output of different ASR systems for the same

An End-to-End Discriminative Approach to Machine Translation

The Prague Bulletin of Mathematical Linguistics NUMBER 96 OCTOBER Ncode: an Open Source Bilingual N-gram SMT Toolkit

Automatic Speech Recognition and Hybrid Machine Translation for High-Quality Closed-Captioning and Subtitling for Video Broadcast

Learning Translations of Named-Entity Phrases from Parallel Corpora

The Prague Bulletin of Mathematical Linguistics NUMBER 93 JANUARY Training Phrase-Based Machine Translation Models on the Cloud

The Impact of Morphological Errors in Phrase-based Statistical Machine Translation from English and German into Swedish

Computer Aided Translation

Training and evaluation of POS taggers on the French MULTITAG corpus

Language and Computation

SYSTRAN 混 合 策 略 汉 英 和 英 汉 机 器 翻 译 系 统

Turker-Assisted Paraphrasing for English-Arabic Machine Translation

NLP Programming Tutorial 5 - Part of Speech Tagging with Hidden Markov Models

Tagging with Hidden Markov Models

Evaluation of speech technologies

Multi-Engine Machine Translation by Recursive Sentence Decomposition

THUTR: A Translation Retrieval System

Hybrid Machine Translation Guided by a Rule Based System

Tuning Methods in Statistical Machine Translation

Language Model of Parsing and Decoding

Big Data and Scripting. (lecture, computer science, bachelor/master/phd)

Introduction. Philipp Koehn. 28 January 2016

Empirical Machine Translation and its Evaluation

SWIFT Aligner, A Multifunctional Tool for Parallel Corpora: Visualization, Word Alignment, and (Morpho)-Syntactic Cross-Language Transfer

Automatic Generation of Bid Phrases for Online Advertising

Future Entreprise. Jean-Dominique Meunier NEM Executive Director Nov. 23, 2009 FIA Stockholm

Learning with Big Data I: Divide and Conquer. Chris Dyer. lti. 25 July 2013 LxMLS. lti

Overview of MT techniques. Malek Boualem (FT)

UNSUPERVISED MORPHOLOGICAL SEGMENTATION FOR STATISTICAL MACHINE TRANSLATION

Machine Translation and the Translator

AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS

Outline. NP-completeness. When is a problem easy? When is a problem hard? Today. Euler Circuits

Building a Web-based parallel corpus and filtering out machinetranslated

Online free translation services

Machine Learning for natural language processing

Factored bilingual n-gram language models for statistical machine translation

Assessments; Optional module 1 vert assessments. Own test on the perfect tense

Dublin City University at CLEF 2004: Experiments with the ImageCLEF St Andrew s Collection

Course: Model, Learning, and Inference: Lecture 5

ISA OR NOT ISA: THE INTERLINGUAL DILEMMA FOR MACHINE TRANSLATION

GCE EXAMINERS' REPORTS. FRENCH AS/Advanced

Factored Translation Models

Appraise: an Open-Source Toolkit for Manual Evaluation of MT Output

Best practices for efficient HPC performance with large models

Spot me if you can: Uncovering spoken phrases in encrypted VoIP conversations

Mrs. Thompson French 4 / UMSL French thompsondella@rockwood.k12.mo.us Web page:

Engineering Political Science

Protein Protein Interaction Networks

Factored Markov Translation with Robust Modeling

CS 534: Computer Vision 3D Model-based recognition

Lesson 1. Je suis béni... (Disc 1, Track 1) Oui, she is magnificent! J aime ma maman beaucoup. Je suis béni to have une maman magnifique!!

CINDOR Conceptual Interlingua Document Retrieval: TREC-8 Evaluation.

The CMU Syntax-Augmented Machine Translation System: SAMT on Hadoop with N-best alignments

Data Mining Practical Machine Learning Tools and Techniques

Transcription:

Statistical NLP Spring 2008 Lecture 11: Word Alignment Dan Klein UC Berkeley Machine Translation: Examples 1

Machine Translation Madame la présidente, votre présidence de cette institution a été marquante. Mrs Fontaine, your presidency of this institution has been outstanding. Madam President, president of this house has been discoveries. Madam President, your presidency of this institution has been impressive. Je vais maintenant m'exprimer brièvement en irlandais. I shall now speak briefly in Irish. I will now speak briefly in Ireland. I will now speak briefly in Irish. Nous trouvons en vous un président tel que nous le souhaitions. We think that you are the type of president that we want. We are in you a president as the wanted. We are in you a president as we the wanted. Levels of Transfer Semantic Composition Interlingua Semantic Decomposition (Vauquois triangle) Semantic Analysis Semantic Structure Semantic Transfer Semantic Structure Semantic Generation Syntactic Analysis Syntactic Structure Syntactic Transfer Syntactic Structure Syntactic Generation Morphological Analysis Word Structure Source Text Direct Word Structure Target Text Morphological Generation 2

General Approaches Rule-based approaches Expert system-like rewrite systems Interlingua methods (analyze and generate) Lexicons come from humans Can be very fast, and can accumulate a lot of knowledge over time (e.g. Systran) Statistical approaches Word-to-word translation Phrase-based translation Syntax-based translation (tree-to-tree, tree-to-string) Trained on parallel corpora Usually noisy-channel (at least in spirit) MT System Components Language Model Translation Model source P(e) e channel P(f e) f best e decoder observed f argmax P(e f) = argmax P(f e)p(e) e e 3

Today The components of a simple MT system You already know about the LM Word-alignment based TMs IBM models 1 and 2, HMM model A simple decoder Next few classes More complex word-level and phrase-level TMs Tree-to-tree and tree-to-string TMs More sophisticated decoders Word Alignment x What is the anticipated cost of collecting fees under the new proposal? En vertu des nouvelles propositions, quel est le coût prévu de perception des droits? What is the anticipated cost of collecting fees under the new proposal? z En vertu de les nouvelles propositions, quel est le coût prévu de perception de les droits? 4

Unsupervised Word Alignment Input: a bitext: pairs of translated sentences nous acceptons votre opinion. we accept your view. Output: alignments: pairs of translated words When words have unique sources, can represent as a (forward) alignment function a from French to English positions 1-to-Many Alignments 5

Many-to-1 Alignments Many-to-Many Alignments 6

A Word-Level TM? What might a model of P(f e) look like? What can go wrong here? How to estimate this? IBM Model 1 (Brown 93) Alignments: a hidden vector called an alignment specifies which English source is responsible for each French target word. 7

IBM Model 1 Obvious first stab: greedy matchings Better approach: re-estimated generative models Basic idea: pick a source for each word, update cooccurrence statistics, repeat Evaluating TMs How do we measure quality of a word-to-word model? Method 1: use in an end-to-end translation system Hard to measure translation quality Option: human judges Option: reference translations (NIST, BLEU) Option: combinations (HTER) Actually, no one uses word-to-word models alone as TMs Method 2: measure quality of the alignments produced Easy to measure Hard to know what the gold alignments should be Often does not correlate well with translation quality (like perplexity in LMs) 8

Alignment Error Rate Alignment Error Rate = = Sure align. Possible align. = Predicted align. Problems with Model 1 There s a reason they designed models 2-5! Problems: alignments jump around, align everything to rare words Experimental setup: Training data: 1.1M sentences of French-English text, Canadian Hansards Evaluation metric: alignment error Rate (AER) Evaluation data: 447 handaligned sentences 9

Intersected Model 1 Post-intersection: standard practice to train models in each direction then intersect their predictions [Och and Ney, 03] Second model is basically a filter on the first Precision jumps, recall drops End up not guessing hard alignments Model Model 1 E F Model 1 F E Model 1 AND P/R 82/58 85/58 96/46 AER 30.6 28.7 34.8 Joint Training? Overall: Similar high precision to post-intersection But recall is much higher More confident about positing non-null alignments Model Model 1 E F Model 1 F E Model 1 AND Model 1 INT P/R 82/58 85/58 96/46 93/69 AER 30.6 28.7 34.8 19.5 10

Monotonic Translation Japan shaken by two new quakes Le Japon secoué par deux nouveaux séismes Local Order Change Japan is at the junction of four tectonic plates Le Japon est au confluent de quatre plaques tectoniques 11

IBM Model 2 Alignments tend to the diagonal (broadly at least) Other schemes for biasing alignments towards the diagonal: Relative vs absolute alignment Asymmetric distances Learning a full multinomial over distances Example 12

EM for Models 1/2 Model 1 Parameters: Translation probabilities (1+2) Distortion parameters (2 only) Start with uniform, including For each sentence: For each French position j Calculate posterior over English positions (or just use best single alignment) Increment count of word f j with word e i by these amounts Also re-estimate distortion probabilities for model 2 Iterate until convergence Phrase Movement On Tuesday Nov. 4, earthquakes rocked Japan once again Des tremblements de terre ont à nouveau touché le Japon jeudi 4 novembre. 13

Phrase Movement The HMM Model Model 2 preferred global monotonicity We want local monotonicity: Most jumps are small HMM model (Vogel 96) Re-estimate using the forward-backward algorithm Handling nulls requires some care What are we still missing? -2-1 0 1 2 3 14

The HMM Model HMM Examples 15

AER for HMMs Model Model 1 INT HMM E F HMM F E HMM AND HMM INT GIZA M4 AND AER 19.5 11.4 10.8 7.1 4.7 6.9 IBM Models 3/4/5 Mary did not slap the green witch Mary not slap slap slap the green witch n(3 slap) P(NULL) Mary not slap slap slap NULL the green witch t(la the) Mary no daba una botefada a la verde bruja d(j i) Mary no daba una botefada a la bruja verde [from Al-Onaizan and Knight, 1998] 16

Examples: Translation and Fertility Example: Idioms 17

Example: Morphology [Och and Ney 03] Some Results 18

Decoding In these word-to-word models Finding best alignments is easy Finding translations is hard (why?) Bag Generation (Decoding) 19

Bag Generation as a TSP Imagine bag generation with a bigram LM Words are nodes Edge weights are P(w w ) Valid sentences are Hamiltonian paths Not the best news for word-based MT! not it clear is. IBM Decoding as a TSP 20

Decoding, Anyway Simplest possible decoder: Enumerate sentences, score each with TM and LM Greedy decoding: Assign each French word it s most likely English translation Operators: Change a translation Insert a word into the English (zero-fertile French) Remove a word from the English (null-generated French) Swap two adjacent English words Do hill-climbing (or annealing) Greedy Decoding 21

Stack Decoding Stack decoding: Beam search Usually A* estimates for completion cost One stack per candidate sentence length Other methods: Dynamic programming decoders possible if we make assumptions about the set of allowable permutations 22