Multi-Engine Machine Translation by Recursive Sentence Decomposition



Similar documents
Empirical Machine Translation and its Evaluation

Accelerating and Evaluation of Syntactic Parsing in Natural Language Question Answering Systems

Dublin City University at CLEF 2004: Experiments with the ImageCLEF St Andrew s Collection

BILINGUAL TRANSLATION SYSTEM

tance alignment and time information to create confusion networks 1 from the output of different ASR systems for the same

Convergence of Translation Memory and Statistical Machine Translation

Effective Self-Training for Parsing

INF5820 Natural Language Processing - NLP. H2009 Jan Tore Lønning jtl@ifi.uio.no

Machine Learning for natural language processing

SYSTRAN Chinese-English and English-Chinese Hybrid Machine Translation Systems for CWMT2011 SYSTRAN 混 合 策 略 汉 英 和 英 汉 机 器 翻 译 系 CWMT2011 技 术 报 告

Shallow Parsing with Apache UIMA

Question Answering and Multilingual CLEF 2008

Customizing an English-Korean Machine Translation System for Patent Translation *

Introduction. BM1 Advanced Natural Language Processing. Alexander Koller. 17 October 2014

Choosing the best machine translation system to translate a sentence by using only source-language information*

Self-Training for Parsing Learner Text

Language and Computation

Binarizing Syntax Trees to Improve Syntax-Based Machine Translation Accuracy

SYSTRAN 混 合 策 略 汉 英 和 英 汉 机 器 翻 译 系 统

Open Domain Information Extraction. Günter Neumann, DFKI, 2012

Natural Language Processing

Overview of MT techniques. Malek Boualem (FT)

A Flexible Online Server for Machine Translation Evaluation

A Framework-based Online Question Answering System. Oliver Scheuer, Dan Shen, Dietrich Klakow

Statistical Machine Translation: IBM Models 1 and 2

EVALITA 07 parsing task

Statistical Machine Translation

Big Data in Education

Syntax for Statistical Machine Translation

Tagging with Hidden Markov Models

Applying Co-Training Methods to Statistical Parsing. Anoop Sarkar anoop/

Sense-Tagging Verbs in English and Chinese. Hoa Trang Dang

Automatic Speech Recognition and Hybrid Machine Translation for High-Quality Closed-Captioning and Subtitling for Video Broadcast

Mining wisdom. Anders Søgaard

The TCH Machine Translation System for IWSLT 2008

The XMU Phrase-Based Statistical Machine Translation System for IWSLT 2006

LIUM s Statistical Machine Translation System for IWSLT 2010

NLP Programming Tutorial 5 - Part of Speech Tagging with Hidden Markov Models

Scalable Inference and Training of Context-Rich Syntactic Translation Models

Statistical Pattern-Based Machine Translation with Statistical French-English Machine Translation

The Specific Text Analysis Tasks at the Beginning of MDA Life Cycle

A Natural Language Interface for Querying General and Individual Knowledge

Left-Brain/Right-Brain Multi-Document Summarization

Machine Translation. Agenda

Log-Linear Models a.k.a. Logistic Regression, Maximum Entropy Models

Why Evaluation? Machine Translation. Evaluation. Evaluation Metrics. Ten Translations of a Chinese Sentence. How good is a given system?

Towards a RB-SMT Hybrid System for Translating Patent Claims Results and Perspectives

Machine Translation. Why Evaluation? Evaluation. Ten Translations of a Chinese Sentence. Evaluation Metrics. But MT evaluation is a di cult problem!

The Adjunct Network Marketing Model

Strategies for Training Large Scale Neural Network Language Models

An End-to-End Discriminative Approach to Machine Translation

Evalita 09 Parsing Task: constituency parsers and the Penn format for Italian

Phase 2 of the D4 Project. Helmut Schmid and Sabine Schulte im Walde

Appraise: an Open-Source Toolkit for Manual Evaluation of MT Output

Turker-Assisted Paraphrasing for English-Arabic Machine Translation

Language Model of Parsing and Decoding

Adapting General Models to Novel Project Ideas

Machine Translation and the Translator

A User-Based Usability Assessment of Raw Machine Translated Technical Instructions

DEPENDENCY PARSING JOAKIM NIVRE

CS 6740 / INFO Ad-hoc IR. Graduate-level introduction to technologies for the computational treatment of information in humanlanguage

POSBIOTM-NER: A Machine Learning Approach for. Bio-Named Entity Recognition

Translation Solution for

Customer Intentions Analysis of Twitter Based on Semantic Patterns

Why language is hard. And what Linguistics has to say about it. Natalia Silveira Participation code: eagles

Course: Model, Learning, and Inference: Lecture 5

Hybrid Machine Translation Guided by a Rule Based System

NP Subject Detection in Verb-Initial Arabic Clauses

NATURAL LANGUAGE QUERY PROCESSING USING PROBABILISTIC CONTEXT FREE GRAMMAR

WHITE PAPER. Machine Translation of Language for Safety Information Sharing Systems

Chunk Parsing. Steven Bird Ewan Klein Edward Loper. University of Melbourne, AUSTRALIA. University of Edinburgh, UK. University of Pennsylvania, USA

Timeline (1) Text Mining Master TKI. Timeline (2) Timeline (3) Overview. What is Text Mining?

ACCURAT Analysis and Evaluation of Comparable Corpora for Under Resourced Areas of Machine Translation Project no.

Learning Translation Rules from Bilingual English Filipino Corpus

INPUTLOG 6.0 a research tool for logging and analyzing writing process data. Linguistic analysis. Linguistic analysis

SZTE-NLP: Aspect Level Opinion Mining Exploiting Syntactic Cues

Improving Data Driven Part-of-Speech Tagging by Morphologic Knowledge Induction

Factored Translation Models

L130: Chapter 5d. Dr. Shannon Bischoff. Dr. Shannon Bischoff () L130: Chapter 5d 1 / 25

Phrase-Based MT. Machine Translation Lecture 7. Instructor: Chris Callison-Burch TAs: Mitchell Stern, Justin Chiu. Website: mt-class.

Report on the embedding and evaluation of the second MT pilot

CINTIL-PropBank. CINTIL-PropBank Sub-corpus id Sentences Tokens Domain Sentences for regression atsts 779 5,654 Test

A Joint Sequence Translation Model with Integrated Reordering

Domain Classification of Technical Terms Using the Web

Grammars and introduction to machine learning. Computers Playing Jeopardy! Course Stony Brook University

Computer Aided Translation

Statistical Machine Translation Lecture 4. Beyond IBM Model 1 to Phrase-Based Models

Transition-Based Dependency Parsing with Long Distance Collocations

Introduction. Philipp Koehn. 28 January 2016

Natural Language to Relational Query by Using Parsing Compiler

POS Tagsets and POS Tagging. Definition. Tokenization. Tagset Design. Automatic POS Tagging Bigram tagging. Maximum Likelihood Estimation 1 / 23

A Hybrid System for Patent Translation

Systematic Comparison of Professional and Crowdsourced Reference Translations for Machine Translation

Polish - English Statistical Machine Translation of Medical Texts.

Hidden Markov Models Chapter 15

TAPTA: Translation Assistant for Patent Titles and Abstracts

The Transition of Phrase based to Factored based Translation for Tamil language in SMT Systems

Project Management. From industrial perspective. A. Helle M. Herranz. EXPERT Summer School, Pangeanic - BI-Europe

Symbiosis of Evolutionary Techniques and Statistical Natural Language Processing

Leveraging ASEAN Economic Community through Language Translation Services

Transcription:

Multi-Engine Machine Translation by Recursive Sentence Decomposition Bart Mellebeek Karolina Owczarzak Josef van Genabith Andy Way National Centre for Language Technology School of Computing Dublin City University 7th Biennial Conference of the Association for Machine Translation in the Americas

Outline What is Multi-Engine Machine Translation? MEMT with Recursive Sentence Decomposition. Experiments, Results and Analysis.

Outline What is Multi-Engine Machine Translation? MEMT with Recursive Sentence Decomposition. Experiments, Results and Analysis.

Multi-Engine Machine Translation (MEMT) Use several MT engines for same input. Combine outputs into consensus translation. Why? Three heads are better than one (Frederking and Nirenburg, 1994) Errors committed by a system are independent of errors commited by other systems. Other domains: Speech Recognition, Text Categorization, POS tagging. The units for comparison are not defined a priori for MT.

Different Approaches to MEMT Previous Approaches Translate entire input sentence individual MT system does not improve. Find consensus through output alignment. Our Approach Translate chunks in context individual MT systems can improve. Prepare input for processing. References Bangalore et al., IEEE 2001. Nomoto, ACL 2004. Jayaraman and Lavie, EAMT 2005. van Zaanen and Somers, MT Summit 2005 Matusov et al., EACL 2006

Different Approaches to MEMT Previous Approaches Translate entire input sentence individual MT system does not improve. Find consensus through output alignment. References Bangalore et al., IEEE 2001. Nomoto, ACL 2004. Jayaraman and Lavie, EAMT 2005. van Zaanen and Somers, MT Summit 2005 Matusov et al., EACL 2006 Our Approach Translate chunks in context individual MT systems can improve. Prepare input for processing. What do we have in common? Majority voting. Language models. Confidence score.

Outline What is Multi-Engine Machine Translation? MEMT with Recursive Sentence Decomposition. Experiments, Results and Analysis.

Overview Architecture Input Decomposition C1... CM E1... EN 1. Decomposition.... C1_1 C1_N CM_1 CM_N 2. Translation. 3. Selection. C1_best Selection... CM_best 4. Composition. Composition Output

Decomposition of Input into Optimal Chunks Input = syntactic parse of string. Decompose into pivot and satellites. Pivot = nucleus (+ additional material). Satellite = argument/adjunct chunk Chunk SAT 1... SAT l pivot SAT l+1... SAT l+r

Example of decomposition The chairman, a long-time rival of Bill, likes fast and confidential deals. S NP-SBJ VP NP, NP, VBZ NP DT the NN chairman DT a, NP JJ NN long-time rival IN of, likes PP NP NNP NNP Bill JJ fast ADJP CC JJ and confidential NNS deals

Example of decomposition [The chairman, a long-time rival of Bill,] ARG1 [likes] pivot [fast and confidential deals] ARG2. S SAT 1 The chairman, a long-time rival of Bill pivot likes SAT 2 fast and confidential deals

Skeletons and Substitution Variables [SAT 1 ]... [SAT l ] pivot [SAT l+1 ]... [SAT l+r ] [V SAT1 ]... [V SATl ] pivot [V SATl+1 ]... [V SATl+r ] V SATi = a simpler string substituting SAT i Reduce complexity original constituents better translation pivot. Track location satellites in target. Substitution Variables: static vs. dynamic Example [The chairman, a long-time rival of Bill,] SAT1 confidential deals] SAT2. [likes] pivot [fast and

Skeletons and Substitution Variables [SAT 1 ]... [SAT l ] pivot [SAT l+1 ]... [SAT l+r ] [V SAT1 ]... [V SATl ] pivot [V SATl+1 ]... [V SATl+r ] V SATi = a simpler string substituting SAT i Reduce complexity original constituents better translation pivot. Track location satellites in target. Substitution Variables: static vs. dynamic Example [The chairman, a long-time rival of Bill,] SAT1 and confidential deals] SAT2. [likes] pivot [fast

Skeletons and Substitution Variables [SAT 1 ]... [SAT l ] pivot [SAT l+1 ]... [SAT l+r ] [V SAT1 ]... [V SATl ] pivot [V SATl+1 ]... [V SATl+r ] V SATi = a simpler string substituting SAT i Reduce complexity original constituents better translation pivot. Track location satellites in target. Substitution Variables: static vs. dynamic Example [The chairman] VSAT1 [likes] pivot [deals] VSAT2.

Skeletons and Substitution Variables [SAT 1 ]... [SAT l ] pivot [SAT l+1 ]... [SAT l+r ] [V SAT1 ]... [V SATl ] pivot [V SATl+1 ]... [V SATl+r ] V SATi = a simpler string substituting SAT i Reduce complexity original constituents better translation pivot. Track location satellites in target. Substitution Variables: static vs. dynamic Example [The boy] VSAT1 [likes] pivot [books] VSAT2.

Translation of Input Chunks If complexity SAT i not OK recursion. If complexity SAT i OK translate. Embed SAT i in context for optimal translation. Context: static vs. dynamic. Example [The chairman, a long-time rival of Bill,] SAT1 confidential deals] SAT2. [likes] pivot [fast and

Translation of Input Chunks If complexity SAT i not OK recursion. If complexity SAT i OK translate. Embed SAT i in context for optimal translation. Context: static vs. dynamic. Example [The chairman, a long-time rival of Bill,] SAT1 [likes] pivot [fast and confidential deals] SAT2.

Translation of Input Chunks If complexity SAT i not OK recursion. If complexity SAT i OK translate. Embed SAT i in context for optimal translation. Context: static vs. dynamic. Example [The chairman likes] Context [fast and confidential deals] SAT2.

Translation of Input Chunks If complexity SAT i not OK recursion. If complexity SAT i OK translate. Embed SAT i in context for optimal translation. Context: static vs. dynamic. Example [The boy sees] Context [fast and confidential deals] SAT2.

Selection of Best Output Chunk The selection of the best translation Ci best for each input chunk C i is based on 3 heuristics: 1. Majority Voting: identical translations are better. 2. Language Modeling: 213M words, trigram model with Kneser-Ney smoothing, SRI. 3. Confidence Score: choose overall best engine if no clear winner.

Composition of Output S SAT 1... SAT i... SAT l pivot SAT l+1... SAT l+r SAT i1... SAT il pivot i SAT il+1... SAT j... SAT il+r SAT j1... SAT jl pivot j SAT jl+1... SAT jl+r Recursive procedure. Only syntactically simple chunks sent to MT engine. Substitution variables track location of translation chunks in target. Therefore composing translation is straightforward.

A Worked Example The chairman a long-time rival of Bill likes fast and confidential deals

A Worked Example The chairman a long-time rival of Bill likes fast and confidential deals

A Worked Example The chairman a long-time rival of Bill likes fast and confidential deals

A Worked Example The chairman a long-time rival of Bill likes fast and confidential deals una largo - vez rival de Bill (-33.77) un rival de largo plazo de Bill (-23.41) un rival antiguo de Bill (-22.60)

A Worked Example The chairman a long-time rival of Bill likes fast and confidential deals una largo - vez rival de Bill (-33.77) un rival de largo plazo de Bill (-23.41) un rival antiguo de Bill (-22.60) un rival antiguo de Bill

A Worked Example The chairman a long-time rival of Bill likes fast and confidential deals una largo - vez rival de Bill le gustan (-33.77) (-10.94) un rival de largo plazo de Bill tiene de gusto (-23.41) (-16.41) un rival antiguo de Bill quiere (-22.60) (-9.73) un rival antiguo de Bill

A Worked Example The chairman a long-time rival of Bill likes fast and confidential deals una largo - vez rival de Bill le gustan (-33.77) (-10.94) un rival de largo plazo de Bill tiene de gusto (-23.41) (-16.41) un rival antiguo de Bill quiere (-22.60) (-9.73) un rival antiguo de Bill quiere

A Worked Example The chairman a long-time rival of Bill likes fast and confidential deals una largo - vez rival de Bill le gustan los los tratos rápidos y confidenciales (-33.77) (-10.94) (-28.13) un rival de largo plazo de Bill tiene de gusto repartos rápidos y confidenciales (-23.41) (-16.41) (-22.16) un rival antiguo de Bill quiere los tratos rápidos y confidenciales (-22.60) (-9.73) (-23.12) un rival antiguo de Bill quiere

A Worked Example The chairman a long-time rival of Bill likes fast and confidential deals una largo - vez rival de Bill le gustan los los tratos rápidos y confidenciales (-33.77) (-10.94) (-28.13) un rival de largo plazo de Bill tiene de gusto repartos rápidos y confidenciales (-23.41) (-16.41) (-22.16) un rival antiguo de Bill quiere los tratos rápidos y confidenciales (-22.60) (-9.73) (-23.12) un rival antiguo de Bill quiere repartos rápidos y confidenciales

A Worked Example The chairman a long-time rival of Bill likes fast and confidential deals una largo - vez rival de Bill le gustan los los tratos rápidos y confidenciales (-33.77) (-10.94) (-28.13) un rival de largo plazo de Bill tiene de gusto repartos rápidos y confidenciales (-23.41) (-16.41) (-22.16) un rival antiguo de Bill quiere los tratos rápidos y confidenciales (-22.60) (-9.73) (-23.12) un rival antiguo de Bill quiere repartos rápidos y confidenciales

A Worked Example Our approach is not limited to a blind combination of previously produced output chunks. An individual MT system might improve its own translation, which can contribute to a better MEMT score. Source The chairman, a long-time rival of Bill, likes fast and confidential deals. Baseline MT, rival de largo plazo de Bill, gustos ayuna y los repartos confidenciales. TransBooster, un rival de largo plazo de Bill, tiene gusto de repartos rápidos y confidenciales.

A Worked Example Our approach is not limited to a blind combination of previously produced output chunks. An individual MT system might improve its own translation, which can contribute to a better MEMT score. Source The chairman, a long-time rival of Bill, likes fast and confidential deals. Baseline MT, rival de largo plazo de Bill, gustos ayuna y los repartos confidenciales. TransBooster, un rival de largo plazo de Bill, tiene gusto de repartos rápidos y confidenciales.

A Worked Example Our approach is not limited to a blind combination of previously produced output chunks. An individual MT system might improve its own translation, which can contribute to a better MEMT score. Source The chairman, a long-time rival of Bill, likes fast and confidential deals. Baseline MT, rival de largo plazo de Bill, gustos ayuna y los repartos confidenciales. TransBooster, un rival de largo plazo de Bill, tiene gusto de repartos rápidos y confidenciales.

Outline What is Multi-Engine Machine Translation? MEMT with Recursive Sentence Decomposition. Experiments, Results and Analysis.

Experimental Setup English Spanish. 800-sentence test set Penn-II treebank. One set of 800 reference translations. Baseline systems: LogoMedia, Systran, SDL. Three different syntactic analyses as input. 1. Original Penn-II structure. 2. (Charniak, 2000) 3. (Bikel, 2002)

Results: Relative Score MEMT wrt Baseline Systems MEMT1 BLEU NIST GTM LogoMedia 104.9 104.8 103.1 Systran 109.7 107.1 104.4 SDL 108.4 105.5 102.4 MEMT2 BLEU NIST GTM LogoMedia 102.1 103.5 102.0 Systran 106.8 105.8 103.4 SDL 105.6 104.2 101.4 MEMT3 BLEU NIST GTM LogoMedia 101.2 103.2 101.8 Systran 105.8 105.5 103.2 SDL 104.6 103.9 101.2 MEMT1: relative BLEU increase of 8.4%-9.7%. MEMT2: relative BLEU increase of 2.1%-6.8%. MEMT3: relative BLEU increase of 1.2%-5.8%.

Conclusions Novel approach to MEMT. Input preparation instead of output alignment. Recursive sentence decomposition. Individual MT systems can improve their output before Multi-Engine selection.

Future Research Experiment with a variety of language models (Nomoto, 2004). Experiment with similarity measure. Implement word graph-based consensus at level of output chunks.

Questions? Thanks for your attention.! "#$ %&