7-2 Speech-to-Speech Translation System Field Experiments in All Over Japan
|
|
|
- Angela Adele Cooper
- 10 years ago
- Views:
Transcription
1 7-2 Speech-to-Speech Translation System Field Experiments in All Over Japan We explain field experiments conducted during the 2009 fiscal year in five areas of Japan. We also show the experiments of evaluation and data selection method from speech translation field data. The data selection method selects useful data from filed data by using a development data set. According to the experimental results, the proposed data selection method gives the improvement of the speech-to-speech translation systems. Keywords Speech translation field experiment, Statistical machine translation, Speech translation system, Speech translation field data 1 Speech translation field experiments In this paper we first explain the speech translation field experiments conducted in 2009, and then the method for improving the speech translation system using the data set collected in the field experiments. Aiming at the significant enhancement of the translation accuracy of automatic speech translation technology and the early realization of a service that utilizes this technology for foreign visitors to Japan, the Ministry of Internal Affairs and Communications conducted field experiments, titled Field Testing of Automatic Speech Translation Technology and Its Contribution to Local Tourism (total works budget of 985 million yen), contracting them to privately-owned corporations. The field experiments were performed with about 1,700 terminals in about 370 tourist facilities in five areas, as shown in Fig. 1, for the four languages of Japanese, English, Chinese, and Korean. About 200,000 accesses were recorded in the period of the experiments. This was the world s first large-scale field experiment conducted under practical conditions. The National Institute of Information and Communications Technology provided its speech translation technology to all of the corporations that were contracted by the Ministry, and offered full cooperation for the development and operation of the experiment system and data analysis. 2 System configuration Figure 2 shows an outline of the configuration of the field experiment system developed by each project area. The speech translation terminal consists of a smart phone and a laptop PC, and 300 to 500 terminals were installed in each area. A speech given to the terminal is sampled at 16 khz and sent in an ADPCM format to a speech translation server. The speech translation server consists of a speech recognition server, a machine translation server, and a speech synthesis server, which are all available for every language. The translation result is sent to the terminal in the form of text or a synthesized speech. Also, the speech input, result of speech recognition, and translation result are saved, together with the date, terminal ID, and specified language, in a log file in the system. 213
2 Fig.1 Outline of fi eld experiments in 5 regions Fig.2 System confi guration of speech translation fi eld experiments 2.1 Outline of speech recognition system The speech recognition system used in this project consists of a front end unit and decoder unit. A particle filter in the front end unit successively estimates the noise strength, which changes with time, to suppress non-stationary noise. For the decoder unit, a Hidden Markov Model (HMM) is used as an acoustic model, and a multi-class composite N-gram, an extended version of word class N-gram, is used as the language model. These are used for two-path speech recognition. In the first path, the acoustic model and 2-gram language model are used to create a word graph. In the second path, a trigram language model is used for word lattice rescoring and to look for a recognition result. The acoustic model is trained by using a 400-hour speech corpus of about 4,500 adult and elder persons. It was expected that the speech recognition system was to be used in a 214 Journal of the National Institute of Information and Communications Technology Vol. 59 Nos. 3/4 2012
3 Fig.3 Process fl ow of machine translations 2.2 Overview of machine translation system Figure 3 shows details of the processing of the machine translation unit in the speech translation system that consists of speech recognition, machine translation and speech synthesis units. The machine translation unit consists mostly of a statistical machine translation unit and two translation memories. For the statistical machine translation system, a phrasebased statistical machine translation framework was used. In this framework, the probability of a word sequence ( ) translated in a target language from its source word sequence ( ) is calculated from the following equation. noisy environment, e.g. outdoors, for the field experiments. We therefore estimated the acoustic model using training data in which automobile noise and other noises recorded in various places, such as streets and railway stations, were randomly superimposed at an S/N ratio of db. We then estimated multi-class composite 2-gram and multi-class 3-gram language models using a text corpus of about 740,000 sentences collected from travel conversations. The lexicon size was about 50,000 words Customization of language model for each area We customized the language model for each area, using proper nouns and regional expressions (dialects) specific to the area. The proper nouns (about 5,000 words) used in these areas were categorized by place name, facility name, and others and the language probability of typical words in the basic dictionary was assigned to each category. Also the regional expressions (about 3,000 sentences) were used for the training of the word N-gram and were combined linearly with the basic language model for the adaptation of the model. Here, represents a candidate translation of. h i is a feature function obtained from a learning corpus. There are eight feature functions, including translation probabilities (translation models) of a word or phrase from target language to source language or from source language to target language, and language models of target languages. i and M are a weight of each feature function and the number of feature functions (8), respectively. Assuming that the denominator of Equation (1) is constant, we derive a translation result ê from the following Equation (2). We used the MOSES tool kit and SRILM tool kit for the learning of the translation models and language models. In the field experiments, the following two kinds of data were collected prior to the experiments in each area. -Proper nouns: Several thousand regional nouns. Their translations (in English, Chinese and Korean) and their categories were also collected. -Regional expressions: Several thousand regional expressions (texts), collected through 215
4 prior interviews in the places where each terminal was to be installed. Their translations (in English, Chinese and Korean) were also collected. The data also included expressions that are necessary for the business of stores where terminals were installed. The proper nouns are used in the translations by replacing them with tokens of the corresponding category in advance as proposed by the reference. In the learning of the models, not only the BTEC corpus but also regional expressions are used. First, the abovementioned tool kits are used for the learning of the feature function for each of the BTEC corpus and regional expressions, and the two feature functions are linearly combined: Here, is the eight feature functions of Equation (1) obtained from the BTEC corpus, and is the eight feature functions obtained from the regional expressions. is the eight feature functions obtained by the linear combination of these two. For the weight μ, we will explain in Subsection 4.1. The regional expressions can be used in another way. Namely, they are incorporated into the corpus, which is then used for the learning of a model. However if there are additional regional expressions, the model learning has to be made again with all the data. On the other hand, with the method of Equation (3), one can make the model learning only with the small-size data of the regional expressions, which provides for easy maintenance. The regional expressions are also used in the translation memories. Translation memory 1 in Fig. 3 uses the BTEC corpus and translation memory 2 uses the regional expressions collected from each area. 3 Use of data collected in field experiment In our proposed method, we use the results of back translations from translated texts to the source language in order to choose the data for model adaptation. In Subsection 3.1, we explain supervised filtering which is based on the proposed method or two conventional methods. In Subsection 3.2 we describe a supervised filtering method, which is based on conventional methods. In Subsection 3.3, we explain how the selected data is used for model adaptation. 3.1 Unsupervised filtering Filtering by normalized translation score In a conventional method that uses normalized translation scores, the following formula is used to calculate the normalized translation score ( ) of data, and if it is larger than a threshold, the data is used for model adaptation. Here, represents the number of words in the translation results and shows the translation probability calculated by Equation (1) Filtering by source language perplexity In another conventional method that uses source language perplexity, the following formula is used to calculate the perplexity ( ) of the data, and if it is larger than a certain threshold, the data is used for model adaptation. Here, represents the number of words in the input sentence and shows the perplexity of the sentence given by the language model of the source language. In this paper, we use the language model for which the learning is made with the source language side of the BTEC corpus in Section Proposed method In our proposed method, the result of a automatically translated text from the source language to the target language is automatically back-translated to the source language. Then, using the result ( ) of the speech recognition 216 Journal of the National Institute of Information and Communications Technology Vol. 59 Nos. 3/4 2012
5 of the source as a reference translation, an automatic evaluation score for the result ( ) of the back-translation is calculated. We use the following PER (position independent word error rate) as the automatic translation evaluation value: Here, is the word error rate calculated from an evaluation of the speech recognition ignoring the word order. For the evaluation of machine translations, the BLEU score is often used as automatic evaluation score. However if BLEU, which should be applied to speech, is applied to the present cases, the score is 0 in most cases and we cannot rate the translation. This is a problem of mismatching between the translation quality and the evaluation method, and we employ the PER which has a relatively higher compatibility between them. 3.2 Supervised filtering Supervised filtering is a method of selecting high-quality translation through the automatic translation evaluation of each sentence in the results of machine translations. For the evaluation, manually-made reference translations are used. Since the aim of the unsupervised adaptation that we use in this paper is adaptation without using manually-made reference translations, the use of these reference translations for filtering is not realistic. In this paper, however, we also conducted supervised filtering to perform additional experiments to the previous study and to compare the performance of the supervised filtering with that of the unsupervised filtering. In the previous study, BLEU was used for automatic translation evaluation. In this paper, on the other hand, PER ( ) calculated from the following formula is used. 3.3 Adaptation method The selected data is used in the following manner, together with the regional expression data mentioned in Section 2. Step 1 The obtained data and the regional expressions mentioned in Section 2 are combined into an adaptation corpus. Step 2 The corpus obtained in Step 1 is used for the learning of the eight feature functions ( ) of Equation (1). Step 3 The model ( ) for which the learning is made with the BTEC corpus is linearly combined with to obtain an adaptation model ( ). The value of the weight the Subsection Experiments will be discussed in Next we examine how to improve the performance of the speech translation system by using actual data collected in the field experiments. In ordinary studies on speech translation systems, the data is manually annotated (transcribed and translated) and the arranged data is used for the re-learning of the system. This method is very effective but requires a great deal of cost and time for the data annotation. To circumvent this problem, we considered using the data without manual annotation. 4.1 Experiment conditions In the preliminary experiment, the subjective evaluation of the translation quality was worst in Hokkaido and highest in Kyushu. We hence used the data obtained in Hokkaido and Kyushu. We focused on translations from Japanese to English since the needs for translation in this direction was high in every area and since we could collect the largest volume of data for this translation direction. The Japanese-English BTEC corpus used for model learning contains 691,829 sentences, and the regional expressions obtained in Hokkaido and those in Kyushu were 3,000 sentences and 5,095 sentences respectively. For the evaluation of the machine translations, we used 100 sentences randomly picked 217
6 up from each region s data. The perplexity of the test set of Hokkaido in the language model for which the learning is performed using only the Japanese BTEC corpus was and that of Kyushu was On the other hand, the perplexity of the test set of Hokkaido with the baseline language model obtained by Equation (3) was and that of Kyushu was We thus see that the test set of Hokkaido was more difficult than that of Kyushu. In the preliminary experiments and the present experiments, we made subjective evaluations of the translation quality using five grades: S (Perfect), A (Correct), B (Fair), C (Acceptable), and D (Nonsense). In the evaluation shown in Subsection 4.2, we employed text input without using speech recognition for the machine translation of a test sentence. The values of the eight weights λ i of the feature functions in Equation (2) were determined by MERT (Minimum Error Rate Training) with a BTEC development set of 500 sentences and were used for all the experiments. In the same manner, the weight μ in Equations (3) and (8) was set to 0.9 according to the preliminary experiment using the BTEC development set. These values were the same as those used in the field experiments. The reasons that we chose these values in the field experiments were as follows. -Development set of actual data cannot be sufficiently obtained before the field experiments. -The speech translation system is used differently depending on when it is used. For example, the system was used during the Snow Festival. Therefore, the parameter tuning which is based on a small-size development set collected in a certain time period may cause over-adaptation. -The parameter tuning based on the BTEC development set can maintain at least the performance of the speech domain of the BTEC. For optimal parameter setting and field data obtained in the entire period of the field experiments should be used for random sampling of development sets from the data, and the weights λ i and μ need to be tuned individually by the development sets. In this experiment, however, we used the same setting as that used in the field experiments since we needed perform experiments under a practical situation where our proposed method will be used. 4.2 Experiment results Table 1 shows the evaluation results of the case where the field data obtained was all used Table 1 Project Area Evaluation results of supervised/unsupervised adaptation (with no data fi ltering) System Type Additional Field Data Ratio (%) Transcription Translation Size (# of sentences) S S, A S, A, B S, A, B, C Baseline N/A N/A Baseline + unannotated data ASR MT Hokkaido Baseline + unannotated data 1 Manual MT Baseline + unannotated data 2 (Upper bound) Manual Manual Kyushu Baseline N/A N/A Baseline + unannotated data ASR MT Baseline + unannotated data 1 Manual MT BL + unannotated data 2 (Upper bound) Manual Manual Baseline with 3000 regional expressions N/A N/A Journal of the National Institute of Information and Communications Technology Vol. 59 Nos. 3/4 2012
7 without data filtering. For each area, the first line of the table shows the baseline and the second is the result of unsupervised adaptation using all the usable field data. The third line shows the results of the case where the transcription of input speech was made manually and machine translation was utilized to use the target language data in the adaptation, i.e. when partially-supervised adaptation is made. The fourth line presents the results of supervised adaptation, where the transcription and translation were both made manually. This result of supervised adaptation gives the upper bound of the performance improvement. The data size varies even in the same area depending on the conditions, because data was not used in the adaptation if output could not be obtained in the speech recognition or machine translation. As mentioned earlier, Kyushu s data contains more regional expressions than Hokkaido s. To find an influence of the number of regional expressions, we show in the last line of Table 1 the results of the case where the number of regional expressions in Kyushu was adjusted by random sampling to 3,000 sentences, the same number as in Hokkaido. We conducted the random sampling three times and averaged the results of the subjective evaluation. The white cells in the table indicate higher performance than the baseline, the gray cells indicate equivalent performance with the baseline, and the dark gray cells indicate lower performance than the baseline *. From Table 1 we see no improvements in the unsupervised adaptation. The partially-supervised adaptation showed performance improvements in one area but performance degradation in the other. On the other hand, the supervised adaptation had performance improvements in both areas. We see from these results that, if we use the field data without filtering it, there could be performance degradation, in particular in the unsupervised adaptation. The filtering of the data should therefore be employed to prevent this degradation and the effect of our proposed method is shown in Table 2. We finally compared the baseline data of the two regions. In Kyushu s data, the performance degraded when the number of regional expressions was decreased to 3,000. However it was still much higher than that of Hokkaido s data which contains the same number of regional expressions. We could therefore conclude that Kyushu s test set was easier to translate than Hokkaido s test set. This difference could be attributed to the fact that, in the field experiments in Kyushu, the terminals were placed at airports and tourist spots with guides under relatively controlled conditions. Table 2 shows the evaluation results obtained by conventional methods and our proposed method, where the speech translation field data is filtered for unsupervised adaptation in the presence of speech recognition errors and machine translation errors. The word error rate of the speech recognition system is 29.9% in Hokkaido and 20.3% in Kyushu. For the proposed method, we set the filtering threshold to be 0.1, 0.2, or 0.4. To make a fair comparison, the thresholds in the conventional methods were adjusted in such a way that the adjusted threshold would give the same number of sentences as that obtained with the thresholds of 0.1, 0.2, or 0.4 in the proposed method. Let us first compare the data quantities when the threshold is set to 0.1 in the proposed method. The filtering of Hokkaido s data gives 1,244 sentences (12.7% of entire data obtained in Hokkaido), while the filtering of Kyushu s data gives 4,560 sentences (46.9% of entire data obtained in Kyushu.) This difference was caused by the following: -In comparison to Hokkaido, Kyushu s data had a low word error rate and hence the speech recognition errors had little influence on the translation quality. As a result, the back-translated texts and the input texts were in relatively good agreement. -In comparison to Hokkaido, Kyushu s The same colors are used in Tables 2 and
8 data contained a lot of easy-to-translate data and was hence translated relatively correctly from source to target language as well as from target to source language. As a result, the back translated texts and the input texts were in relatively good agreement. Next we focus on the data set of Hokkaido. With the proposed method, the system performance improved from the baseline in almost all cases except the case of the threshold equal to 0.1, where the performance degraded. The performance improved also when the normalized translation score (S trans ) or source language perplexity (S pp ) was used, although the improvement from the proposed method was greater. Comparing the proposed method (S src_per ) to the supervised filtering (S tgt_per ), we see that the proposed method improved performance even though it was unsupervised, more than the supervised filtering method. With the condition S src_per < = 0.2, the total ratio of S, A, B, and C was higher than the corresponding Upper bound shown in Table 1. This indicated that when there is no translation error or speech recognition error the translation quality could be improved by eliminating the training sentences that are not included in the domain. This could accidentally happen in the case of S src_per < = 0.2. We considered that even the Upper bound in Table 1 could be exceeded Table 2 Evaluation results of unsupervised adaptation (with data fi ltering) Project Area Additional field data Ratio (%) Filtering function Size (# of sentences) S S, A S, A, B S, A, B, C N/A (Baseline) S trans (eq. 4) S pp (eq. 5) Hokkaido S src_per < = 0.1 (eq. 6) S src_per < = 0.2 (eq. 6) S src_per < = 0.4 (eq. 6) S tgt_per (eq. 7) N/A (Baseline) S trans (eq. 4) S pp (eq. 5) Kyushu S src_per < = 0.1 (eq. 6) S src_per < = 0.2 (eq. 6) S src_per < = 0.4 (eq. 6) S tgt_per (eq. 7) Journal of the National Institute of Information and Communications Technology Vol. 59 Nos. 3/4 2012
9 by using the method of for data selection Detailed analysis of Kyushu s data set Table 3 shows the results of the proposed method where the field data was used only for the language model adaptation and/or the translation model adaptation. We see from the table that, in Hokkaido s data set, the performance was most improved when the adaptation was made for both the language and translation models and the improvement was small when the adaptation was conducted only for the language model or for the translation model. In Kyushu s data set, on the other hand, the adaptation of only the language model degraded performance, while the adaptation of only the translation model improved performance. In actual operations, using a development data set to determine a model which is suitable for adaptation would prevent performance degradation and improve performance. 5 Summary We proposed a machine translation adaptation method using field data from speech translations, which contains speech recognition data and machine translation data. In our proposed method, the results of the machine translation were translated back to the source language, and the back-translated text was compared with the results of the speech recognition. The data was employed for adaptation if its input text (results of the speech recognition) and the back-translated text are close to each other. We conducted experiments using the data Table 3 Project Area Hokkaido Kyushu Evaluation results of each model adaptation (with data fi ltering) Filtering function Additional field data Ratio (%) Used for LM training Used for TM training S S, A S, A, B S, A, B, C N/A (Baseline) No No S src_per < = 0.1 Yes Yes S src_per < = 0.2 Yes Yes S src_per < = 0.4 Yes Yes S src_per < = 0.1 Yes No S src_per < = 0.2 Yes No S src_per < = 0.4 Yes No S src_per < = 0.1 No Yes S src_per < = 0.2 No Yes S src_per < = 0.4 No Yes N/A (Baseline) No No S src_per < = 0.1 Yes Yes S src_per < = 0.2 Yes Yes S src_per < = 0.4 Yes Yes S src_per < = 0.1 Yes No S src_per < = 0.2 Yes No S src_per < = 0.4 Yes No S src_per < = 0.1 No Yes S src_per < = 0.2 No Yes S src_per < = 0.4 No Yes
10 that was obtained in the speech translation field experiments conducted in As a result we found that the proposed method improved the translation performance for the data obtained in the Hokkaido area whose baseline performance was low. On the other hand the baseline performance for the data obtained in Kyushu was relatively high and the translation performance was quite degraded by the present method. However if we made the adaptation not for the language model, but only for the translation model, the performance was improved to a certain degree. In actual operations, it is expected that system performance would be improved by the proposed method without the manual transcription or manual creation of parallel texts, although we need to prepare a development set to determine the threshold for data filtering and to determine a model to which the adaptation can be made. The result of a questionnaire survey of the field experiments shows that the performance of current speech translation systems is not high enough to satisfy everyone. The present method has a function of adjusting the probability of each model using field data but cannot handle unknown words contained in the field data. For further improvements of the system, we need to incorporate a mechanism for acquiring proper nouns automatically from the Web into the speech translation system and to continue to collect field data. References 1 H. Kawai, R. Isitani, K. Yasuda, E. Sumita, M. Uchiyama, S. Matsuda, Y. Ashikari, and S. Nakamura, An overview of a nation-wide field experiment of speech-to-speech translation in fiscal year 2009, Proc. of 2010 Autumn Meeting of the ASJ, pp , M. Fujimoto and S. Nakamura, A non-stationary noise suppression method based on particle fi ltering and polyak averaging, The IEICE Transactions on Information and Systems, Vol. E89-D, No. 3, pp , H. Yamamoto and Y. Sagisaka, Multi-Class Composite N-gram Language Model Based on Connection Direction, The IEICE Transactions on Information and Systems, Vol.J83-D-II, No.11, pp , P. Koehn, F. J. Och, and D. Marcu, Statistical Phrase-Based Translation, Proc. of Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics (HLT-NAACL), pp , P. Koehn, H. Hoang, A. Birch, C. Callison-Burch, M. Federico, N. Bertoldi, B. Cowan, W. Shen, C. Moran, R. Zens, C. Dyer, O. Bojar, A. Constantin, and E. Herbst, Moses: Open source toolkit for statistical machine translation, Proceedings of the 45th Annual Meeting of the Association for Computational Linguistics Companion Volume Proceedings of the Demo and Poster Sessions, pp , Association for Computational Linguistics, June A. Stolcke, SRILM-an extensible language modeling toolkit, Proceedings of the International Conference on Spoken Language Processing, pp , H. Okuma, H. Yamamoto, and E. Sumita, Introducint a translation dictionary into phrasebased smt, The IEICE Transactions on Information and Systems, vol. 91-D, no. 7, pp , G. Kikui, S. Yamamoto, T. Takezawa, and E. Sumita, Comparative study on corpora for speech translation, IEEE Transactions on Audio, Speech and Language Processing, vol. 14(5), pp , N. Ueffi ng, G. Haffari, and A. Sarkar, Semi-supervised model adaptation for statistical machine translation, Machine Translation, Vol. 21, No. 2, pp , K. Yasuda, R. Zhang, H. Yamamoto, and E. Sumita, Method of selecting training data to build a compact 222 Journal of the National Institute of Information and Communications Technology Vol. 59 Nos. 3/4 2012
11 and efficient translation model, Proceedings of the Third International Joint Conference on Natural Language Processing, pp , K. Papineni, S. Roukos, T. Ward, and W.-J. Zhu, Bleu: a method for automatic evaluation of machine translation, Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL), pp , N. Bach, R. Hsiao, M. Eck, P. Charoenpornsawat, S. Vogel, T. Schultz, I. Lane, A. Waibel, and A. W. Black, Incremental adaptation of speech-to-speech translation, Proceedings of NAACL HLT 2009, pp , F. Och, Minimum error rate training in statistical machine translation, Proc. of 41st Annual Meeting of the Association for Computational Linguistics (ACL), pp , K. Yasuda, H. Yamamoto, and E. Sumita, Training set selection for building compact and effi cient language models, The IEICE Transactions on Information and Systems, Vol. 92-D, No. 3, pp , (Accepted June 14, 2012) YASUDA Keiji, Dr. Eng. Senior Researcher, Multilingual Translation Laboratory, Universal Communication Research Institute Machine Translation, Natural Language Processing MATSUDA Shigeki, Ph.D. Senior Researcher, Spoken Language Communication Laboratory, Universal Communication Research Institute Signal Processing, Speech Recognition 223
THUTR: A Translation Retrieval System
THUTR: A Translation Retrieval System Chunyang Liu, Qi Liu, Yang Liu, and Maosong Sun Department of Computer Science and Technology State Key Lab on Intelligent Technology and Systems National Lab for
The XMU Phrase-Based Statistical Machine Translation System for IWSLT 2006
The XMU Phrase-Based Statistical Machine Translation System for IWSLT 2006 Yidong Chen, Xiaodong Shi Institute of Artificial Intelligence Xiamen University P. R. China November 28, 2006 - Kyoto 13:46 1
Segmentation and Punctuation Prediction in Speech Language Translation Using a Monolingual Translation System
Segmentation and Punctuation Prediction in Speech Language Translation Using a Monolingual Translation System Eunah Cho, Jan Niehues and Alex Waibel International Center for Advanced Communication Technologies
Hybrid Machine Translation Guided by a Rule Based System
Hybrid Machine Translation Guided by a Rule Based System Cristina España-Bonet, Gorka Labaka, Arantza Díaz de Ilarraza, Lluís Màrquez Kepa Sarasola Universitat Politècnica de Catalunya University of the
Building a Web-based parallel corpus and filtering out machinetranslated
Building a Web-based parallel corpus and filtering out machinetranslated text Alexandra Antonova, Alexey Misyurev Yandex 16, Leo Tolstoy St., Moscow, Russia {antonova, misyurev}@yandex-team.ru Abstract
Factored Translation Models
Factored Translation s Philipp Koehn and Hieu Hoang [email protected], [email protected] School of Informatics University of Edinburgh 2 Buccleuch Place, Edinburgh EH8 9LW Scotland, United Kingdom
Collaborative Machine Translation Service for Scientific texts
Collaborative Machine Translation Service for Scientific texts Patrik Lambert [email protected] Jean Senellart Systran SA [email protected] Laurent Romary Humboldt Universität Berlin
The Prague Bulletin of Mathematical Linguistics NUMBER 96 OCTOBER 2011 49 58. Ncode: an Open Source Bilingual N-gram SMT Toolkit
The Prague Bulletin of Mathematical Linguistics NUMBER 96 OCTOBER 2011 49 58 Ncode: an Open Source Bilingual N-gram SMT Toolkit Josep M. Crego a, François Yvon ab, José B. Mariño c c a LIMSI-CNRS, BP 133,
The Impact of Morphological Errors in Phrase-based Statistical Machine Translation from English and German into Swedish
The Impact of Morphological Errors in Phrase-based Statistical Machine Translation from English and German into Swedish Oscar Täckström Swedish Institute of Computer Science SE-16429, Kista, Sweden [email protected]
UM-Corpus: A Large English-Chinese Parallel Corpus for Statistical Machine Translation
UM-Corpus: A Large English-Chinese Parallel Corpus for Statistical Machine Translation Liang Tian 1, Derek F. Wong 1, Lidia S. Chao 1, Paulo Quaresma 2,3, Francisco Oliveira 1, Yi Lu 1, Shuo Li 1, Yiming
Machine translation techniques for presentation of summaries
Grant Agreement Number: 257528 KHRESMOI www.khresmoi.eu Machine translation techniques for presentation of summaries Deliverable number D4.6 Dissemination level Public Delivery date April 2014 Status Author(s)
OPTIMIZATION OF NEURAL NETWORK LANGUAGE MODELS FOR KEYWORD SEARCH. Ankur Gandhe, Florian Metze, Alex Waibel, Ian Lane
OPTIMIZATION OF NEURAL NETWORK LANGUAGE MODELS FOR KEYWORD SEARCH Ankur Gandhe, Florian Metze, Alex Waibel, Ian Lane Carnegie Mellon University Language Technology Institute {ankurgan,fmetze,ahw,lane}@cs.cmu.edu
Turkish Radiology Dictation System
Turkish Radiology Dictation System Ebru Arısoy, Levent M. Arslan Boaziçi University, Electrical and Electronic Engineering Department, 34342, Bebek, stanbul, Turkey [email protected], [email protected]
Automatic Speech Recognition and Hybrid Machine Translation for High-Quality Closed-Captioning and Subtitling for Video Broadcast
Automatic Speech Recognition and Hybrid Machine Translation for High-Quality Closed-Captioning and Subtitling for Video Broadcast Hassan Sawaf Science Applications International Corporation (SAIC) 7990
Adaptive Development Data Selection for Log-linear Model in Statistical Machine Translation
Adaptive Development Data Selection for Log-linear Model in Statistical Machine Translation Mu Li Microsoft Research Asia [email protected] Dongdong Zhang Microsoft Research Asia [email protected]
An Iteratively-Trained Segmentation-Free Phrase Translation Model for Statistical Machine Translation
An Iteratively-Trained Segmentation-Free Phrase Translation Model for Statistical Machine Translation Robert C. Moore Chris Quirk Microsoft Research Redmond, WA 98052, USA {bobmoore,chrisq}@microsoft.com
SYSTRAN Chinese-English and English-Chinese Hybrid Machine Translation Systems for CWMT2011 SYSTRAN 混 合 策 略 汉 英 和 英 汉 机 器 翻 译 系 CWMT2011 技 术 报 告
SYSTRAN Chinese-English and English-Chinese Hybrid Machine Translation Systems for CWMT2011 Jin Yang and Satoshi Enoue SYSTRAN Software, Inc. 4444 Eastgate Mall, Suite 310 San Diego, CA 92121, USA E-mail:
Human in the Loop Machine Translation of Medical Terminology
Human in the Loop Machine Translation of Medical Terminology by John J. Morgan ARL-MR-0743 April 2010 Approved for public release; distribution unlimited. NOTICES Disclaimers The findings in this report
Factored Markov Translation with Robust Modeling
Factored Markov Translation with Robust Modeling Yang Feng Trevor Cohn Xinkai Du Information Sciences Institue Computing and Information Systems Computer Science Department The University of Melbourne
UEdin: Translating L1 Phrases in L2 Context using Context-Sensitive SMT
UEdin: Translating L1 Phrases in L2 Context using Context-Sensitive SMT Eva Hasler ILCC, School of Informatics University of Edinburgh [email protected] Abstract We describe our systems for the SemEval
Factored bilingual n-gram language models for statistical machine translation
Mach Translat DOI 10.1007/s10590-010-9082-5 Factored bilingual n-gram language models for statistical machine translation Josep M. Crego François Yvon Received: 2 November 2009 / Accepted: 12 June 2010
Automatic Mining of Internet Translation Reference Knowledge Based on Multiple Search Engines
, 22-24 October, 2014, San Francisco, USA Automatic Mining of Internet Translation Reference Knowledge Based on Multiple Search Engines Baosheng Yin, Wei Wang, Ruixue Lu, Yang Yang Abstract With the increasing
An Online Service for SUbtitling by MAchine Translation
SUMAT CIP-ICT-PSP-270919 An Online Service for SUbtitling by MAchine Translation Annual Public Report 2012 Editor(s): Contributor(s): Reviewer(s): Status-Version: Arantza del Pozo Mirjam Sepesy Maucec,
Applying Statistical Post-Editing to. English-to-Korean Rule-based Machine Translation System
Applying Statistical Post-Editing to English-to-Korean Rule-based Machine Translation System Ki-Young Lee and Young-Gil Kim Natural Language Processing Team, Electronics and Telecommunications Research
Systematic Comparison of Professional and Crowdsourced Reference Translations for Machine Translation
Systematic Comparison of Professional and Crowdsourced Reference Translations for Machine Translation Rabih Zbib, Gretchen Markiewicz, Spyros Matsoukas, Richard Schwartz, John Makhoul Raytheon BBN Technologies
Automatic slide assignation for language model adaptation
Automatic slide assignation for language model adaptation Applications of Computational Linguistics Adrià Agustí Martínez Villaronga May 23, 2013 1 Introduction Online multimedia repositories are rapidly
Statistical Machine Translation
Statistical Machine Translation Some of the content of this lecture is taken from previous lectures and presentations given by Philipp Koehn and Andy Way. Dr. Jennifer Foster National Centre for Language
Domain Adaptation for Medical Text Translation Using Web Resources
Domain Adaptation for Medical Text Translation Using Web Resources Yi Lu, Longyue Wang, Derek F. Wong, Lidia S. Chao, Yiming Wang, Francisco Oliveira Natural Language Processing & Portuguese-Chinese Machine
The Prague Bulletin of Mathematical Linguistics NUMBER 93 JANUARY 2010 37 46. Training Phrase-Based Machine Translation Models on the Cloud
The Prague Bulletin of Mathematical Linguistics NUMBER 93 JANUARY 2010 37 46 Training Phrase-Based Machine Translation Models on the Cloud Open Source Machine Translation Toolkit Chaski Qin Gao, Stephan
Translating the Penn Treebank with an Interactive-Predictive MT System
IJCLA VOL. 2, NO. 1 2, JAN-DEC 2011, PP. 225 237 RECEIVED 31/10/10 ACCEPTED 26/11/10 FINAL 11/02/11 Translating the Penn Treebank with an Interactive-Predictive MT System MARTHA ALICIA ROCHA 1 AND JOAN
Customizing an English-Korean Machine Translation System for Patent Translation *
Customizing an English-Korean Machine Translation System for Patent Translation * Sung-Kwon Choi, Young-Gil Kim Natural Language Processing Team, Electronics and Telecommunications Research Institute,
Domain Classification of Technical Terms Using the Web
Systems and Computers in Japan, Vol. 38, No. 14, 2007 Translated from Denshi Joho Tsushin Gakkai Ronbunshi, Vol. J89-D, No. 11, November 2006, pp. 2470 2482 Domain Classification of Technical Terms Using
Turker-Assisted Paraphrasing for English-Arabic Machine Translation
Turker-Assisted Paraphrasing for English-Arabic Machine Translation Michael Denkowski and Hassan Al-Haj and Alon Lavie Language Technologies Institute School of Computer Science Carnegie Mellon University
Generating Training Data for Medical Dictations
Generating Training Data for Medical Dictations Sergey Pakhomov University of Minnesota, MN [email protected] Michael Schonwetter Linguistech Consortium, NJ [email protected] Joan Bachenko
The noisier channel : translation from morphologically complex languages
The noisier channel : translation from morphologically complex languages Christopher J. Dyer Department of Linguistics University of Maryland College Park, MD 20742 [email protected] Abstract This paper
IMPLEMENTING SRI S PASHTO SPEECH-TO-SPEECH TRANSLATION SYSTEM ON A SMART PHONE
IMPLEMENTING SRI S PASHTO SPEECH-TO-SPEECH TRANSLATION SYSTEM ON A SMART PHONE Jing Zheng, Arindam Mandal, Xin Lei 1, Michael Frandsen, Necip Fazil Ayan, Dimitra Vergyri, Wen Wang, Murat Akbacak, Kristin
Jane 2: Open Source Phrase-based and Hierarchical Statistical Machine Translation
Jane 2: Open Source Phrase-based and Hierarchical Statistical Machine Translation Joern Wuebker M at thias Huck Stephan Peitz M al te Nuhn M arkus F reitag Jan-Thorsten Peter Saab M ansour Hermann N e
Appraise: an Open-Source Toolkit for Manual Evaluation of MT Output
Appraise: an Open-Source Toolkit for Manual Evaluation of MT Output Christian Federmann Language Technology Lab, German Research Center for Artificial Intelligence, Stuhlsatzenhausweg 3, D-66123 Saarbrücken,
SOME INSIGHTS FROM TRANSLATING CONVERSATIONAL TELEPHONE SPEECH. Gaurav Kumar, Matt Post, Daniel Povey, Sanjeev Khudanpur
SOME INSIGHTS FROM TRANSLATING CONVERSATIONAL TELEPHONE SPEECH Gaurav Kumar, Matt Post, Daniel Povey, Sanjeev Khudanpur Center for Language and Speech Processing & Human Language Technology Center of Excellence
Cache-based Online Adaptation for Machine Translation Enhanced Computer Assisted Translation
Cache-based Online Adaptation for Machine Translation Enhanced Computer Assisted Translation Nicola Bertoldi Mauro Cettolo Marcello Federico FBK - Fondazione Bruno Kessler via Sommarive 18 38123 Povo,
Chapter 6. Decoding. Statistical Machine Translation
Chapter 6 Decoding Statistical Machine Translation Decoding We have a mathematical model for translation p(e f) Task of decoding: find the translation e best with highest probability Two types of error
Word Completion and Prediction in Hebrew
Experiments with Language Models for בס"ד Word Completion and Prediction in Hebrew 1 Yaakov HaCohen-Kerner, Asaf Applebaum, Jacob Bitterman Department of Computer Science Jerusalem College of Technology
SYSTRAN 混 合 策 略 汉 英 和 英 汉 机 器 翻 译 系 统
SYSTRAN Chinese-English and English-Chinese Hybrid Machine Translation Systems Jin Yang, Satoshi Enoue Jean Senellart, Tristan Croiset SYSTRAN Software, Inc. SYSTRAN SA 9333 Genesee Ave. Suite PL1 La Grande
Structural and Semantic Indexing for Supporting Creation of Multilingual Web Pages
Structural and Semantic Indexing for Supporting Creation of Multilingual Web Pages Hiroshi URAE, Taro TEZUKA, Fuminori KIMURA, and Akira MAEDA Abstract Translating webpages by machine translation is the
The Prague Bulletin of Mathematical Linguistics NUMBER 93 JANUARY 2010 7 16
The Prague Bulletin of Mathematical Linguistics NUMBER 93 JANUARY 21 7 16 A Productivity Test of Statistical Machine Translation Post-Editing in a Typical Localisation Context Mirko Plitt, François Masselot
Strategies for Training Large Scale Neural Network Language Models
Strategies for Training Large Scale Neural Network Language Models Tomáš Mikolov #1, Anoop Deoras 2, Daniel Povey 3, Lukáš Burget #4, Jan Honza Černocký #5 # Brno University of Technology, Speech@FIT,
Micro blogs Oriented Word Segmentation System
Micro blogs Oriented Word Segmentation System Yijia Liu, Meishan Zhang, Wanxiang Che, Ting Liu, Yihe Deng Research Center for Social Computing and Information Retrieval Harbin Institute of Technology,
A GrAF-compliant Indonesian Speech Recognition Web Service on the Language Grid for Transcription Crowdsourcing
A GrAF-compliant Indonesian Speech Recognition Web Service on the Language Grid for Transcription Crowdsourcing LAW VI JEJU 2012 Bayu Distiawan Trisedya & Ruli Manurung Faculty of Computer Science Universitas
Input Support System for Medical Records Created Using a Voice Memo Recorded by a Mobile Device
International Journal of Signal Processing Systems Vol. 3, No. 2, December 2015 Input Support System for Medical Records Created Using a Voice Memo Recorded by a Mobile Device K. Kurumizawa and H. Nishizaki
tance alignment and time information to create confusion networks 1 from the output of different ASR systems for the same
1222 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 16, NO. 7, SEPTEMBER 2008 System Combination for Machine Translation of Spoken and Written Language Evgeny Matusov, Student Member,
TRANSREAD LIVRABLE 3.1 QUALITY CONTROL IN HUMAN TRANSLATIONS: USE CASES AND SPECIFICATIONS. Projet ANR 201 2 CORD 01 5
Projet ANR 201 2 CORD 01 5 TRANSREAD Lecture et interaction bilingues enrichies par les données d'alignement LIVRABLE 3.1 QUALITY CONTROL IN HUMAN TRANSLATIONS: USE CASES AND SPECIFICATIONS Avril 201 4
Hardware Implementation of Probabilistic State Machine for Word Recognition
IJECT Vo l. 4, Is s u e Sp l - 5, Ju l y - Se p t 2013 ISSN : 2230-7109 (Online) ISSN : 2230-9543 (Print) Hardware Implementation of Probabilistic State Machine for Word Recognition 1 Soorya Asokan, 2
The United Nations Parallel Corpus v1.0
The United Nations Parallel Corpus v1.0 Michał Ziemski, Marcin Junczys-Dowmunt, Bruno Pouliquen United Nations, DGACM, New York, United States of America Adam Mickiewicz University, Poznań, Poland World
IBM Research Report. Scaling Shrinkage-Based Language Models
RC24970 (W1004-019) April 6, 2010 Computer Science IBM Research Report Scaling Shrinkage-Based Language Models Stanley F. Chen, Lidia Mangu, Bhuvana Ramabhadran, Ruhi Sarikaya, Abhinav Sethy IBM Research
Improving MT System Using Extracted Parallel Fragments of Text from Comparable Corpora
mproving MT System Using Extracted Parallel Fragments of Text from Comparable Corpora Rajdeep Gupta, Santanu Pal, Sivaji Bandyopadhyay Department of Computer Science & Engineering Jadavpur University Kolata
Convergence of Translation Memory and Statistical Machine Translation
Convergence of Translation Memory and Statistical Machine Translation Philipp Koehn and Jean Senellart 4 November 2010 Progress in Translation Automation 1 Translation Memory (TM) translators store past
THE RWTH ENGLISH LECTURE RECOGNITION SYSTEM
THE RWTH ENGLISH LECTURE RECOGNITION SYSTEM Simon Wiesler 1, Kazuki Irie 2,, Zoltán Tüske 1, Ralf Schlüter 1, Hermann Ney 1,2 1 Human Language Technology and Pattern Recognition, Computer Science Department,
Transcription System Using Automatic Speech Recognition for the Japanese Parliament (Diet)
Proceedings of the Twenty-Fourth Innovative Appications of Artificial Intelligence Conference Transcription System Using Automatic Speech Recognition for the Japanese Parliament (Diet) Tatsuya Kawahara
Oersetter: Frisian-Dutch Statistical Machine Translation
Oersetter: Frisian-Dutch Statistical Machine Translation Maarten van Gompel and Antal van den Bosch, Centre for Language Studies, Radboud University Nijmegen and Anne Dykstra, Fryske Akademy > Abstract
A New Input Method for Human Translators: Integrating Machine Translation Effectively and Imperceptibly
Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI 2015) A New Input Method for Human Translators: Integrating Machine Translation Effectively and Imperceptibly
Wikipedia and Web document based Query Translation and Expansion for Cross-language IR
Wikipedia and Web document based Query Translation and Expansion for Cross-language IR Ling-Xiang Tang 1, Andrew Trotman 2, Shlomo Geva 1, Yue Xu 1 1Faculty of Science and Technology, Queensland University
Network Sensing Network Monitoring and Diagnosis Technologies
Network Sensing Network Monitoring and Diagnosis Technologies Masanobu Morinaga Yuji Nomura Takeshi Otani Motoyuki Kawaba (Manuscript received April 7, 2009) Problems that occur in the information and
An End-to-End Discriminative Approach to Machine Translation
An End-to-End Discriminative Approach to Machine Translation Percy Liang Alexandre Bouchard-Côté Dan Klein Ben Taskar Computer Science Division, EECS Department University of California at Berkeley Berkeley,
Robust Methods for Automatic Transcription and Alignment of Speech Signals
Robust Methods for Automatic Transcription and Alignment of Speech Signals Leif Grönqvist ([email protected]) Course in Speech Recognition January 2. 2004 Contents Contents 1 1 Introduction 2 2 Background
Application of Machine Translation in Localization into Low-Resourced Languages
Application of Machine Translation in Localization into Low-Resourced Languages Raivis Skadiņš 1, Mārcis Pinnis 1, Andrejs Vasiļjevs 1, Inguna Skadiņa 1, Tomas Hudik 2 Tilde 1, Moravia 2 {raivis.skadins;marcis.pinnis;andrejs;inguna.skadina}@tilde.lv,
Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information
Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information Satoshi Sekine Computer Science Department New York University [email protected] Kapil Dalwani Computer Science Department
Natural Language to Relational Query by Using Parsing Compiler
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 3, March 2015,
Chinese-Japanese Machine Translation Exploiting Chinese Characters
Chinese-Japanese Machine Translation Exploiting Chinese Characters CHENHUI CHU, TOSHIAKI NAKAZAWA, DAISUKE KAWAHARA, and SADAO KUROHASHI, Kyoto University The Chinese and Japanese languages share Chinese
Audio Coding Algorithm for One-Segment Broadcasting
Audio Coding Algorithm for One-Segment Broadcasting V Masanao Suzuki V Yasuji Ota V Takashi Itoh (Manuscript received November 29, 2007) With the recent progress in coding technologies, a more efficient
Statistical Machine Translation prototype using UN parallel documents
Proceedings of the 16th EAMT Conference, 28-30 May 2012, Trento, Italy Statistical Machine Translation prototype using UN parallel documents Bruno Pouliquen, Christophe Mazenc World Intellectual Property
