Evaluation of Acoustic Model and Unsupervised Speaker Adaptation for Child Speech Recognition in Real Environment

Size: px
Start display at page:

Download "Evaluation of Acoustic Model and Unsupervised Speaker Adaptation for Child Speech Recognition in Real Environment"

Transcription

1 Vol. 47 No. 7 July 2006, 71.1% 23.9% 59, % 1.7% 0.5% Evaluation of Acoustic Model and Unsupervised Speaker Adaptation for Child Speech Recognition in Real Environment Mitsuru Samejima, Randy Gomez, Akinobu Lee,, Hiroshi Saruwatari and Kiyohiro Shikano Child s utterance has totally different property from adult s speech, not only by their acoustic property, but by their incorrect pronunciation and totally ill-formed speaking style. The rapid physiological changes during the growth also prevent accurate speech recognition using a single model. However, collection of child s read speech is difficult in natural, since forcing them to read a sentence precisely will make the utterances far from spontaneous one. In this research, we evaluated acoustic models and an unsupervised adaptation method based on a large number of real spontaneous child speech automatically collected through an actual spoken dialogue system. Acoustic model trained by an actual spontaneous speech achieves the word accuracy of 71.1%, which outperforms one trained by read speech by 23.9%. Detailed investigation is carried out for child s ages (infant pupils, lower-grade elementary schoolers and higher-grade elementary schoolers), and accuracy of the infant pupils was greatly improved by using the age-dependent model. Then a speaker clustering method is proposed to perform unsupervised speaker adaptation based on HMM Sufficient Statistics on automatically collected database where no user tag is available. Clustering the 59,966 utterances to 200 speaker clusters, and selecting the neighbor one for each input to construct the adapted model has resulted in a further improvement of recognition accuracy by 1.5% as compared with age-class dependent models. 1. Graduate School of Information Science, Nara Institute of Science and Technology Presently with Graduate School of Engineering, Nagoya Institute of Technology 2295

2 2296 July ) Vocal Tract Length Normalizatioin VTLN 2) 5) HMM 1,100 6) 2) 78 7) 354 8) 9) 3 10) 2 3 HMM ) Fig. 1 Spoken-dialogue information agent Takemarukun.

3 Vol. 47 No Table 1 Breakdown of spontaneous child speech data collected by spoken dialogue system. Table 2 2 Most frequent utterances in each child age classes. 9, , , ,966 1,200 / 11) , , % 59, % 2.3 SI SI JNAS 12) Phonetic Tied- Mixture PTM triphone HMM 1 EM (%) 834 (8.03%) 251 (2.42%) 239 (2.30%) 169 (1.63%) 157 (1.51%) 147 (1.42%) 128 (1.23%) 119 (1.15%) 115 (1.11%) 114 (1.10%) 1,638 (4.25%) 731 (1.90%) 673 (1.75%) 672 (1.74%) 594 (1.54%) 558 (1.45%) 518 (1.35%) 438 (1.14%) 433 (1.12%) 377 (0.98%) 299 (2.44%) 218 (1.78%) 203 (1.65%) 198 (1.61%) 164 (1.34%) 163 (1.33%) 152 (1.24%) 147 (1.20%) 130 (1.06%) 119 (0.97%) Table 3 / 3 Conditions of acoustic analysis. 25 ms, 10 ms 16 khz/16 bit MFCC+ MFCC+ Power 25 EM JNAS CSRC SI

4 2298 July Table 4 Experimental conditions for evaluation of age class dependent models. Julius ver gram 40 k 6 Table 6 Specifications of test set for each age classes. test set perp Table 5 5 Breakdown of acoustic models. JNAS 47 JNAS 51 JNAS 98 CSRC 40,000 SI 15,707 /9 SI 59,966 /34 PTM ,000 CSRC 3,000 5 JNAS JNAS CSRC JNAS MAP CSRC ) SI SI JNAS EM PTM ,000 CSRC 3, gram Julius 14) ver ) Web N-gram 15) 4 3-gram 6, Fig. 2 Results of children s speech recognition using SI model. 2 % SI 71.1% CSRC 47.2% 23.9% JNAS SI JNAS CSRC SI SI 15% SI CSRC 42.3% 4

5 Vol. 47 No Fig. 3 3 Evaluation of age-dependent models. SI 82.1% 77.6% 53.6% SI 4. Maximum likelihood Linear Regression MLLR 16) 10) HMM-stat 4.1 HMM EM Gaussian Mixture Model GMM 1 HMM-stat 17) HMM-stat HMM-stat

6 2300 July ) 2 4 K-means MFCC 19) MFCC step 1 N step 2 step 3 step 4 4 Fig. 4 Automatic speaker clustering. 5 Fig. 5 Distance measuring using vowel segments. step 2 step HMM-stat 17) 2 6 step 1 N step 2 HMM GMM HMM step 3 GMM M step 4

7 Vol. 47 No Table 7 7 Experimental conditions for evaluation of proposed adaptation method. SI N 200 M Table 8 Selection rate of age-class model for age evaluation data. 6 HMMstat Fig. 6 HMM-statistics speaker adaptation using multiple initial acoustic models and automatic speaker clustering. 89.5% 3.3% 3.5% 10.5% 96.8% 17.3% 0.0% 0.0% 79.3% step 5 step 3 M HMM HMM-stat 7 59, N = 200 HMM-stat M SI % SI 7 Fig. 7 Results of children s speech recognition using adapted models. 20) 3.4% 2.2% 2.0% SI 5.3 N ,000 N M

8 2302 July Table 9 Word accuracy of adapted model for each cluster granularity % 81.1% 78.0% % 83.7% 78.5% % 84.1% 78.4% % 84.3% 79.6% % 84.2% 79.2% % 83.1% 78.3% 1, % 83.5% 78.6% % 23.9% 53.6% 82.1% 77.6% 59, % 1.7% 0.5% 1) software/ 2) Narayanan, S. and Potamianos, A.: Creating conversational interfaces for children, IEEE Trans. Speech and Audio Processing, Vol.10, No.2, pp (Feb. 2002). 3) Arunachalam, S., Gould, D., Andersen, E., Byrd, D. and Narayanan, S.: Politeness and frustration language in Child-Machine interactions, Proc. EUROSPEECH, pp (Sep. 2001). 4) Potamianos, A., Narayanan, S. and Lee, S.: Automatic speech recognition for children, Proc. EUROSPEECH, pp (Sep. 1997). 5) Wilpon, J.G. and Jacobsen, C.N.: A study of speech recognition for children and elderly, Proc. ICASSP, pp (May 1996). 6) Shobaki, K., Hosom, J.P. and Cole, R.A.: The OGI kid s speech corpus and recognizers, Proc. ICSLP, Vol.4, pp (2000). 7) 2-Q-7, pp (Sep. 2000). 8) Vol.J87-D-II, No.8, pp (2004). 9) Vol.J87-D-II, No.3, pp , ) Yoshizawa, S., Baba, A., Matsunami, K., Mera, Y., Yamada, M. and Shikano, K.: Unsupervised speaker adaptation based on Sufficient HMM Statistics of selected speakers, Proc. ICASSP, pp (2001). 11) 2004-SLP-53-9, pp (2004). 12) Itou, K., Yamamoto, M., Takeda, K., Takezawa, T., Matsuoka, T., Kobayashi, T., shikano, K. and Itahashi, S.: JNAS: Japanese speech corpus for large vocabulary continuous speech recognition research, The Journal of the Acoustical Society of Japan (E), Vol.20, pp (1999). 13) 2002

9 Vol. 47 No SLP-48-1 (2003). 14) Kawahara, T., Lee, A., Kobayashi, T., Takeda, K., Minematsu, N., Sagayama, S., ITou, A., Ito, K., Yamamoto, M., Yamada, A., Utsuro, T. and Shikano, K.: Free software toolkit for japanese large vocabulary continuous speech recognition, Proc. ICSLP, Ob(16)- V-07, pp.iv (2000). 15) N-gram 2003-SLP-45-13, pp (2003). 16) Leggetter, C.J. and Woodland, C.: Maximum likelihood linear regression for speaker adaptation of continuous density hidden markov models, Computer Speech and Language, Vol.9, pp (1995). 17) Gomez, R., Lee, A., Saruwatari, H. and Shikano, K.: Unsupervised Speaker Adaptation Based on HMM Sufficient Statistics Using Multiple Acoustic Models Under Noisy Environment, Proc. Acoustic Sciety of Japan, pp (2004). 18) Nguyen, L., Matsoukas, S., Davenport, J., Kubala, F. and Schwartz, R.: Progress in transcription of Broadcast News using Byblos, Speech Communication, Vol.38, pp (2002). 19) S83-48, pp (1983). 20) 2004-SLP , pp (2004). ( ) ( ) IP FAX Randy Gomez 1998 Mindanao State University-lligan Institute of Technology 2002 University of New South Wales-Electrical Engineering 2003 RA IEEE IS VR IEEE

10 2304 July ATR 4 NTT IEEE SP 1990 Senior Award VR IEEE ISCA VR

SOME ASPECTS OF ASR TRANSCRIPTION BASED UNSUPERVISED SPEAKER ADAPTATION FOR HMM SPEECH SYNTHESIS

SOME ASPECTS OF ASR TRANSCRIPTION BASED UNSUPERVISED SPEAKER ADAPTATION FOR HMM SPEECH SYNTHESIS SOME ASPECTS OF ASR TRANSCRIPTION BASED UNSUPERVISED SPEAKER ADAPTATION FOR HMM SPEECH SYNTHESIS Bálint Tóth, Tibor Fegyó, Géza Németh Department of Telecommunications and Media Informatics Budapest University

More information

AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS

AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS PIERRE LANCHANTIN, ANDREW C. MORRIS, XAVIER RODET, CHRISTOPHE VEAUX Very high quality text-to-speech synthesis can be achieved by unit selection

More information

Turkish Radiology Dictation System

Turkish Radiology Dictation System Turkish Radiology Dictation System Ebru Arısoy, Levent M. Arslan Boaziçi University, Electrical and Electronic Engineering Department, 34342, Bebek, stanbul, Turkey arisoyeb@boun.edu.tr, arslanle@boun.edu.tr

More information

Membering T M : A Conference Call Service with Speaker-Independent Name Dialing on AIN

Membering T M : A Conference Call Service with Speaker-Independent Name Dialing on AIN PAGE 30 Membering T M : A Conference Call Service with Speaker-Independent Name Dialing on AIN Sung-Joon Park, Kyung-Ae Jang, Jae-In Kim, Myoung-Wan Koo, Chu-Shik Jhon Service Development Laboratory, KT,

More information

ADVANCES IN ARABIC BROADCAST NEWS TRANSCRIPTION AT RWTH. David Rybach, Stefan Hahn, Christian Gollan, Ralf Schlüter, Hermann Ney

ADVANCES IN ARABIC BROADCAST NEWS TRANSCRIPTION AT RWTH. David Rybach, Stefan Hahn, Christian Gollan, Ralf Schlüter, Hermann Ney ADVANCES IN ARABIC BROADCAST NEWS TRANSCRIPTION AT RWTH David Rybach, Stefan Hahn, Christian Gollan, Ralf Schlüter, Hermann Ney Human Language Technology and Pattern Recognition Computer Science Department,

More information

THE RWTH ENGLISH LECTURE RECOGNITION SYSTEM

THE RWTH ENGLISH LECTURE RECOGNITION SYSTEM THE RWTH ENGLISH LECTURE RECOGNITION SYSTEM Simon Wiesler 1, Kazuki Irie 2,, Zoltán Tüske 1, Ralf Schlüter 1, Hermann Ney 1,2 1 Human Language Technology and Pattern Recognition, Computer Science Department,

More information

EVALUATION OF AUTOMATIC TRANSCRIPTION SYSTEMS FOR THE JUDICIAL DOMAIN

EVALUATION OF AUTOMATIC TRANSCRIPTION SYSTEMS FOR THE JUDICIAL DOMAIN EVALUATION OF AUTOMATIC TRANSCRIPTION SYSTEMS FOR THE JUDICIAL DOMAIN J. Lööf (1), D. Falavigna (2),R.Schlüter (1), D. Giuliani (2), R. Gretter (2),H.Ney (1) (1) Computer Science Department, RWTH Aachen

More information

Speech Recognition on Cell Broadband Engine UCRL-PRES-223890

Speech Recognition on Cell Broadband Engine UCRL-PRES-223890 Speech Recognition on Cell Broadband Engine UCRL-PRES-223890 Yang Liu, Holger Jones, John Johnson, Sheila Vaidya (Lawrence Livermore National Laboratory) Michael Perrone, Borivoj Tydlitat, Ashwini Nanda

More information

Recent Development of Open-Source Speech Recognition Engine Julius

Recent Development of Open-Source Speech Recognition Engine Julius Recent Development of Open-Source Speech Recognition Engine Julius Akinobu Lee and Tatsuya Kawahara Nagoya Institute of Technology, Nagoya, Aichi 466-8555, Japan E-mail: ri@nitech.ac.jp Kyoto University,

More information

Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics

Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics Anastasis Kounoudes 1, Anixi Antonakoudi 1, Vasilis Kekatos 2 1 The Philips College, Computing and Information Systems

More information

Generating Training Data for Medical Dictations

Generating Training Data for Medical Dictations Generating Training Data for Medical Dictations Sergey Pakhomov University of Minnesota, MN pakhomov.sergey@mayo.edu Michael Schonwetter Linguistech Consortium, NJ MSchonwetter@qwest.net Joan Bachenko

More information

Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition

Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition , Lisbon Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition Wolfgang Macherey Lars Haferkamp Ralf Schlüter Hermann Ney Human Language Technology

More information

Speech Recognition System of Arabic Alphabet Based on a Telephony Arabic Corpus

Speech Recognition System of Arabic Alphabet Based on a Telephony Arabic Corpus Speech Recognition System of Arabic Alphabet Based on a Telephony Arabic Corpus Yousef Ajami Alotaibi 1, Mansour Alghamdi 2, and Fahad Alotaiby 3 1 Computer Engineering Department, King Saud University,

More information

Speech: A Challenge to Digital Signal Processing Technology for Human-to-Computer Interaction

Speech: A Challenge to Digital Signal Processing Technology for Human-to-Computer Interaction : A Challenge to Digital Signal Processing Technology for Human-to-Computer Interaction Urmila Shrawankar Dept. of Information Technology Govt. Polytechnic, Nagpur Institute Sadar, Nagpur 440001 (INDIA)

More information

Connected Digits Recognition Task: ISTC CNR Comparison of Open Source Tools

Connected Digits Recognition Task: ISTC CNR Comparison of Open Source Tools Connected Digits Recognition Task: ISTC CNR Comparison of Open Source Tools Piero Cosi, Mauro Nicolao Istituto di Scienze e Tecnologie della Cognizione, C.N.R. via Martiri della libertà, 2 35137 Padova

More information

Input Support System for Medical Records Created Using a Voice Memo Recorded by a Mobile Device

Input Support System for Medical Records Created Using a Voice Memo Recorded by a Mobile Device International Journal of Signal Processing Systems Vol. 3, No. 2, December 2015 Input Support System for Medical Records Created Using a Voice Memo Recorded by a Mobile Device K. Kurumizawa and H. Nishizaki

More information

Robust Methods for Automatic Transcription and Alignment of Speech Signals

Robust Methods for Automatic Transcription and Alignment of Speech Signals Robust Methods for Automatic Transcription and Alignment of Speech Signals Leif Grönqvist (lgr@msi.vxu.se) Course in Speech Recognition January 2. 2004 Contents Contents 1 1 Introduction 2 2 Background

More information

The effect of mismatched recording conditions on human and automatic speaker recognition in forensic applications

The effect of mismatched recording conditions on human and automatic speaker recognition in forensic applications Forensic Science International 146S (2004) S95 S99 www.elsevier.com/locate/forsciint The effect of mismatched recording conditions on human and automatic speaker recognition in forensic applications A.

More information

Spot me if you can: Uncovering spoken phrases in encrypted VoIP conversations

Spot me if you can: Uncovering spoken phrases in encrypted VoIP conversations Spot me if you can: Uncovering spoken phrases in encrypted VoIP conversations C. Wright, L. Ballard, S. Coull, F. Monrose, G. Masson Talk held by Goran Doychev Selected Topics in Information Security and

More information

ADAPTIVE AND DISCRIMINATIVE MODELING FOR IMPROVED MISPRONUNCIATION DETECTION. Horacio Franco, Luciana Ferrer, and Harry Bratt

ADAPTIVE AND DISCRIMINATIVE MODELING FOR IMPROVED MISPRONUNCIATION DETECTION. Horacio Franco, Luciana Ferrer, and Harry Bratt ADAPTIVE AND DISCRIMINATIVE MODELING FOR IMPROVED MISPRONUNCIATION DETECTION Horacio Franco, Luciana Ferrer, and Harry Bratt Speech Technology and Research Laboratory, SRI International, Menlo Park, CA

More information

SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA

SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA Nitesh Kumar Chaudhary 1 and Shraddha Srivastav 2 1 Department of Electronics & Communication Engineering, LNMIIT, Jaipur, India 2 Bharti School Of Telecommunication,

More information

How to Improve the Sound Quality of Your Microphone

How to Improve the Sound Quality of Your Microphone An Extension to the Sammon Mapping for the Robust Visualization of Speaker Dependencies Andreas Maier, Julian Exner, Stefan Steidl, Anton Batliner, Tino Haderlein, and Elmar Nöth Universität Erlangen-Nürnberg,

More information

CallSurf - Automatic transcription, indexing and structuration of call center conversational speech for knowledge extraction and query by content

CallSurf - Automatic transcription, indexing and structuration of call center conversational speech for knowledge extraction and query by content CallSurf - Automatic transcription, indexing and structuration of call center conversational speech for knowledge extraction and query by content Martine Garnier-Rizet 1, Gilles Adda 2, Frederik Cailliau

More information

Transcription System Using Automatic Speech Recognition for the Japanese Parliament (Diet)

Transcription System Using Automatic Speech Recognition for the Japanese Parliament (Diet) Proceedings of the Twenty-Fourth Innovative Appications of Artificial Intelligence Conference Transcription System Using Automatic Speech Recognition for the Japanese Parliament (Diet) Tatsuya Kawahara

More information

MISSING FEATURE RECONSTRUCTION AND ACOUSTIC MODEL ADAPTATION COMBINED FOR LARGE VOCABULARY CONTINUOUS SPEECH RECOGNITION

MISSING FEATURE RECONSTRUCTION AND ACOUSTIC MODEL ADAPTATION COMBINED FOR LARGE VOCABULARY CONTINUOUS SPEECH RECOGNITION MISSING FEATURE RECONSTRUCTION AND ACOUSTIC MODEL ADAPTATION COMBINED FOR LARGE VOCABULARY CONTINUOUS SPEECH RECOGNITION Ulpu Remes, Kalle J. Palomäki, and Mikko Kurimo Adaptive Informatics Research Centre,

More information

Establishing the Uniqueness of the Human Voice for Security Applications

Establishing the Uniqueness of the Human Voice for Security Applications Proceedings of Student/Faculty Research Day, CSIS, Pace University, May 7th, 2004 Establishing the Uniqueness of the Human Voice for Security Applications Naresh P. Trilok, Sung-Hyuk Cha, and Charles C.

More information

Online Diarization of Telephone Conversations

Online Diarization of Telephone Conversations Odyssey 2 The Speaker and Language Recognition Workshop 28 June July 2, Brno, Czech Republic Online Diarization of Telephone Conversations Oshry Ben-Harush, Itshak Lapidot, Hugo Guterman Department of

More information

Building A Vocabulary Self-Learning Speech Recognition System

Building A Vocabulary Self-Learning Speech Recognition System INTERSPEECH 2014 Building A Vocabulary Self-Learning Speech Recognition System Long Qin 1, Alexander Rudnicky 2 1 M*Modal, 1710 Murray Ave, Pittsburgh, PA, USA 2 Carnegie Mellon University, 5000 Forbes

More information

Emotion Detection from Speech

Emotion Detection from Speech Emotion Detection from Speech 1. Introduction Although emotion detection from speech is a relatively new field of research, it has many potential applications. In human-computer or human-human interaction

More information

Hardware Implementation of Probabilistic State Machine for Word Recognition

Hardware Implementation of Probabilistic State Machine for Word Recognition IJECT Vo l. 4, Is s u e Sp l - 5, Ju l y - Se p t 2013 ISSN : 2230-7109 (Online) ISSN : 2230-9543 (Print) Hardware Implementation of Probabilistic State Machine for Word Recognition 1 Soorya Asokan, 2

More information

The LIMSI RT-04 BN Arabic System

The LIMSI RT-04 BN Arabic System The LIMSI RT-04 BN Arabic System Abdel. Messaoudi, Lori Lamel and Jean-Luc Gauvain Spoken Language Processing Group LIMSI-CNRS, BP 133 91403 Orsay cedex, FRANCE {abdel,gauvain,lamel}@limsi.fr ABSTRACT

More information

Automatic slide assignation for language model adaptation

Automatic slide assignation for language model adaptation Automatic slide assignation for language model adaptation Applications of Computational Linguistics Adrià Agustí Martínez Villaronga May 23, 2013 1 Introduction Online multimedia repositories are rapidly

More information

Automatic speech recognition using articulatory features from subject-independent acoustic-to-articulatory inversion

Automatic speech recognition using articulatory features from subject-independent acoustic-to-articulatory inversion Automatic speech recognition using articulatory features from subject-independent acoustic-to-articulatory inversion Prasanta Kumar Ghosh a) and Shrikanth Narayanan Signal Analysis and Interpretation Laboratory,

More information

ADAPTIVE AND ONLINE SPEAKER DIARIZATION FOR MEETING DATA. Multimedia Communications Department, EURECOM, Sophia Antipolis, France 2

ADAPTIVE AND ONLINE SPEAKER DIARIZATION FOR MEETING DATA. Multimedia Communications Department, EURECOM, Sophia Antipolis, France 2 3rd European ignal Processing Conference (EUIPCO) ADAPTIVE AND ONLINE PEAKER DIARIZATION FOR MEETING DATA Giovanni oldi, Christophe Beaugeant and Nicholas Evans Multimedia Communications Department, EURECOM,

More information

TOOLS FOR RESEARCH AND EDUCATION IN SPEECH SCIENCE

TOOLS FOR RESEARCH AND EDUCATION IN SPEECH SCIENCE TOOLS FOR RESEARCH AND EDUCATION IN SPEECH SCIENCE Ronald A. Cole Center for Spoken Language Understanding, Univ. of Colorado, Boulder ABSTRACT The Center for Spoken Language Understanding (CSLU) provides

More information

Automatic Classification and Transcription of Telephone Speech in Radio Broadcast Data

Automatic Classification and Transcription of Telephone Speech in Radio Broadcast Data Automatic Classification and Transcription of Telephone Speech in Radio Broadcast Data Alberto Abad, Hugo Meinedo, and João Neto L 2 F - Spoken Language Systems Lab INESC-ID / IST, Lisboa, Portugal {Alberto.Abad,Hugo.Meinedo,Joao.Neto}@l2f.inesc-id.pt

More information

A Computer Program for Pronunciation Training and Assessment in Japanese Language Classrooms Experimental Use of Pronunciation Check

A Computer Program for Pronunciation Training and Assessment in Japanese Language Classrooms Experimental Use of Pronunciation Check A Computer Program for Pronunciation Training and Assessment in Japanese Language Classrooms Experimental Use of Pronunciation Check 5 CHIHARU TSURUTANI, Griffith University, Australia 10 E-learning is

More information

7-2 Speech-to-Speech Translation System Field Experiments in All Over Japan

7-2 Speech-to-Speech Translation System Field Experiments in All Over Japan 7-2 Speech-to-Speech Translation System Field Experiments in All Over Japan We explain field experiments conducted during the 2009 fiscal year in five areas of Japan. We also show the experiments of evaluation

More information

An Arabic Text-To-Speech System Based on Artificial Neural Networks

An Arabic Text-To-Speech System Based on Artificial Neural Networks Journal of Computer Science 5 (3): 207-213, 2009 ISSN 1549-3636 2009 Science Publications An Arabic Text-To-Speech System Based on Artificial Neural Networks Ghadeer Al-Said and Moussa Abdallah Department

More information

Improving Automatic Forced Alignment for Dysarthric Speech Transcription

Improving Automatic Forced Alignment for Dysarthric Speech Transcription Improving Automatic Forced Alignment for Dysarthric Speech Transcription Yu Ting Yeung 2, Ka Ho Wong 1, Helen Meng 1,2 1 Human-Computer Communications Laboratory, Department of Systems Engineering and

More information

PodCastle: A Spoken Document Retrieval Service Improved by Anonymous User Contributions

PodCastle: A Spoken Document Retrieval Service Improved by Anonymous User Contributions PACLIC 24 Proceedings 3 PodCastle: A Spoken Document Retrieval Service Improved by Anonymous User Contributions Masataka Goto and Jun Ogata National Institute of Advanced Industrial Science and Technology

More information

Speech recognition for human computer interaction

Speech recognition for human computer interaction Speech recognition for human computer interaction Ubiquitous computing seminar FS2014 Student report Niklas Hofmann ETH Zurich hofmannn@student.ethz.ch ABSTRACT The widespread usage of small mobile devices

More information

KL-DIVERGENCE REGULARIZED DEEP NEURAL NETWORK ADAPTATION FOR IMPROVED LARGE VOCABULARY SPEECH RECOGNITION

KL-DIVERGENCE REGULARIZED DEEP NEURAL NETWORK ADAPTATION FOR IMPROVED LARGE VOCABULARY SPEECH RECOGNITION KL-DIVERGENCE REGULARIZED DEEP NEURAL NETWORK ADAPTATION FOR IMPROVED LARGE VOCABULARY SPEECH RECOGNITION Dong Yu 1, Kaisheng Yao 2, Hang Su 3,4, Gang Li 3, Frank Seide 3 1 Microsoft Research, Redmond,

More information

Ericsson T18s Voice Dialing Simulator

Ericsson T18s Voice Dialing Simulator Ericsson T18s Voice Dialing Simulator Mauricio Aracena Kovacevic, Anna Dehlbom, Jakob Ekeberg, Guillaume Gariazzo, Eric Lästh and Vanessa Troncoso Dept. of Signals Sensors and Systems Royal Institute of

More information

BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION

BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION P. Vanroose Katholieke Universiteit Leuven, div. ESAT/PSI Kasteelpark Arenberg 10, B 3001 Heverlee, Belgium Peter.Vanroose@esat.kuleuven.ac.be

More information

Creating a Virtual Neighbor

Creating a Virtual Neighbor Creating a Virtual Neighbor Carina Corbin, Fabrizio Morbini, and David Traum Abstract We present the first version of our Virtual Neighbor, who can talk with users about people employed in the same institution.

More information

3-2 Language Learning in the 21st Century

3-2 Language Learning in the 21st Century 3-2 Language Learning in the 21st Century Advanced Telecommunications Research Institute International With the globalization of human activities, it has become increasingly important to improve one s

More information

Evaluating grapheme-to-phoneme converters in automatic speech recognition context

Evaluating grapheme-to-phoneme converters in automatic speech recognition context Evaluating grapheme-to-phoneme converters in automatic speech recognition context Denis Jouvet, Dominique Fohr, Irina Illina To cite this version: Denis Jouvet, Dominique Fohr, Irina Illina. Evaluating

More information

Slovak Automatic Transcription and Dictation System for the Judicial Domain

Slovak Automatic Transcription and Dictation System for the Judicial Domain Slovak Automatic Transcription and Dictation System for the Judicial Domain Milan Rusko 1, Jozef Juhár 2, Marian Trnka 1, Ján Staš 2, Sakhia Darjaa 1, Daniel Hládek 2, Miloš Cerňak 1, Marek Papco 2, Róbert

More information

Experiments with Signal-Driven Symbolic Prosody for Statistical Parametric Speech Synthesis

Experiments with Signal-Driven Symbolic Prosody for Statistical Parametric Speech Synthesis Experiments with Signal-Driven Symbolic Prosody for Statistical Parametric Speech Synthesis Fabio Tesser, Giacomo Sommavilla, Giulio Paci, Piero Cosi Institute of Cognitive Sciences and Technologies, National

More information

Lecture 12: An Overview of Speech Recognition

Lecture 12: An Overview of Speech Recognition Lecture : An Overview of peech Recognition. Introduction We can classify speech recognition tasks and systems along a set of dimensions that produce various tradeoffs in applicability and robustness. Isolated

More information

Automatic Transcription of Conversational Telephone Speech

Automatic Transcription of Conversational Telephone Speech IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 13, NO. 6, NOVEMBER 2005 1173 Automatic Transcription of Conversational Telephone Speech Thomas Hain, Member, IEEE, Philip C. Woodland, Member, IEEE,

More information

Speech Recognition for Real-time Closed Captioning

Speech Recognition for Real-time Closed Captioning Speech for Real-time Closed Captioning Closed captioning to convey the speech of TV programs by text is becoming a useful means of providing information for elderly people and the hearing impaired, and

More information

Detecting and Correcting Transcription Discrepancies between Thai Parliament Meeting Speech Utterances and their Official Meeting Reports

Detecting and Correcting Transcription Discrepancies between Thai Parliament Meeting Speech Utterances and their Official Meeting Reports Detecting and Correcting Transcription Discrepancies between Thai Parliament Meeting Speech Utterances and their Official Meeting Reports Natnarong. Puangsri, Atiwong. Suchato, Proadpran. Punyabukkana,

More information

Effects of Pronunciation Practice System Based on Personalized CG Animations of Mouth Movement Model

Effects of Pronunciation Practice System Based on Personalized CG Animations of Mouth Movement Model Effects of Pronunciation Practice System Based on Personalized CG Animations of Mouth Movement Model Kohei Arai 1 Graduate School of Science and Engineering Saga University Saga City, Japan Mariko Oda

More information

OPTIMIZATION ON VIETNAMESE LARGE VOCABULARY SPEECH RECOGNITION. Ngoc Thang Vu, Tanja Schultz

OPTIMIZATION ON VIETNAMESE LARGE VOCABULARY SPEECH RECOGNITION. Ngoc Thang Vu, Tanja Schultz OPTIMIZATION ON VIETNAMESE LARGE VOCABULARY SPEECH RECOGNITION Ngoc Thang Vu, Tanja Schultz Cognitive Lab (CSL), Institute for Anthropomatics, Karlsruhe Institute of Technology (KIT) ABSTRACT This paper

More information

CALL-TYPE CLASSIFICATION AND UNSUPERVISED TRAINING FOR THE CALL CENTER DOMAIN

CALL-TYPE CLASSIFICATION AND UNSUPERVISED TRAINING FOR THE CALL CENTER DOMAIN CALL-TYPE CLASSIFICATION AND UNSUPERVISED TRAINING FOR THE CALL CENTER DOMAIN Min Tang, Bryan Pellom, Kadri Hacioglu Center for Spoken Language Research University of Colorado at Boulder Boulder, Colorado

More information

HMM-based Breath and Filled Pauses Elimination in ASR

HMM-based Breath and Filled Pauses Elimination in ASR HMM-based Breath and Filled Pauses Elimination in ASR Piotr Żelasko 1, Tomasz Jadczyk 1,2 and Bartosz Ziółko 1,2 1 Faculty of Computer Science, Electronics and Telecommunications, AGH University of Science

More information

CHANWOO KIM (BIRTH: APR. 9, 1976) Language Technologies Institute School of Computer Science Aug. 8, 2005 present

CHANWOO KIM (BIRTH: APR. 9, 1976) Language Technologies Institute School of Computer Science Aug. 8, 2005 present CHANWOO KIM (BIRTH: APR. 9, 1976) 2602E NSH Carnegie Mellon University 5000 Forbes Avenue Pittsburgh, PA 15213 Phone: +1-412-726-3996 Email: chanwook@cs.cmu.edu RESEARCH INTERESTS Speech recognition system,

More information

Available from Deakin Research Online:

Available from Deakin Research Online: This is the authors final peered reviewed (post print) version of the item published as: Adibi,S 2014, A low overhead scaled equalized harmonic-based voice authentication system, Telematics and informatics,

More information

Automatic Emotion Recognition from Speech

Automatic Emotion Recognition from Speech Automatic Emotion Recognition from Speech A PhD Research Proposal Yazid Attabi and Pierre Dumouchel École de technologie supérieure, Montréal, Canada Centre de recherche informatique de Montréal, Montréal,

More information

Bayesian Network Modeling of Hangul Characters for On-line Handwriting Recognition

Bayesian Network Modeling of Hangul Characters for On-line Handwriting Recognition Bayesian Network Modeling of Hangul haracters for On-line Handwriting Recognition Sung-ung ho and in H. Kim S Div., EES Dept., KAIS, 373- Kusong-dong, Yousong-ku, Daejon, 305-70, KOREA {sjcho, jkim}@ai.kaist.ac.kr

More information

Evaluation of Interactive User Corrections for Lecture Transcription

Evaluation of Interactive User Corrections for Lecture Transcription Evaluation of Interactive User Corrections for Lecture Transcription Henrich Kolkhorst, Kevin Kilgour, Sebastian Stüker, and Alex Waibel International Center for Advanced Communication Technologies InterACT

More information

German Speech Recognition: A Solution for the Analysis and Processing of Lecture Recordings

German Speech Recognition: A Solution for the Analysis and Processing of Lecture Recordings German Speech Recognition: A Solution for the Analysis and Processing of Lecture Recordings Haojin Yang, Christoph Oehlke, Christoph Meinel Hasso Plattner Institut (HPI), University of Potsdam P.O. Box

More information

SPEAKER IDENTITY INDEXING IN AUDIO-VISUAL DOCUMENTS

SPEAKER IDENTITY INDEXING IN AUDIO-VISUAL DOCUMENTS SPEAKER IDENTITY INDEXING IN AUDIO-VISUAL DOCUMENTS Mbarek Charhad, Daniel Moraru, Stéphane Ayache and Georges Quénot CLIPS-IMAG BP 53, 38041 Grenoble cedex 9, France Georges.Quenot@imag.fr ABSTRACT The

More information

Objective Intelligibility Assessment of Text-to-Speech Systems Through Utterance Verification

Objective Intelligibility Assessment of Text-to-Speech Systems Through Utterance Verification Objective Intelligibility Assessment of Text-to-Speech Systems Through Utterance Verification Raphael Ullmann 1,2, Ramya Rasipuram 1, Mathew Magimai.-Doss 1, and Hervé Bourlard 1,2 1 Idiap Research Institute,

More information

Recognition of Emotions in Interactive Voice Response Systems

Recognition of Emotions in Interactive Voice Response Systems Recognition of Emotions in Interactive Voice Response Systems Sherif Yacoub, Steve Simske, Xiaofan Lin, John Burns HP Laboratories Palo Alto HPL-2003-136 July 2 nd, 2003* E-mail: {sherif.yacoub, steven.simske,

More information

Automatic Detection of Laughter and Fillers in Spontaneous Mobile Phone Conversations

Automatic Detection of Laughter and Fillers in Spontaneous Mobile Phone Conversations Automatic Detection of Laughter and Fillers in Spontaneous Mobile Phone Conversations Hugues Salamin, Anna Polychroniou and Alessandro Vinciarelli University of Glasgow - School of computing Science, G128QQ

More information

Estonian Large Vocabulary Speech Recognition System for Radiology

Estonian Large Vocabulary Speech Recognition System for Radiology Estonian Large Vocabulary Speech Recognition System for Radiology Tanel Alumäe, Einar Meister Institute of Cybernetics Tallinn University of Technology, Estonia October 8, 2010 Alumäe, Meister (TUT, Estonia)

More information

AUDIMUS.media: A Broadcast News Speech Recognition System for the European Portuguese Language

AUDIMUS.media: A Broadcast News Speech Recognition System for the European Portuguese Language AUDIMUS.media: A Broadcast News Speech Recognition System for the European Portuguese Language Hugo Meinedo, Diamantino Caseiro, João Neto, and Isabel Trancoso L 2 F Spoken Language Systems Lab INESC-ID

More information

Unsupervised Estimation of Writing Style Models for Improved Unconstrained Off-line Handwriting Recognition

Unsupervised Estimation of Writing Style Models for Improved Unconstrained Off-line Handwriting Recognition Unsupervised Estimation of Writing Style Models for Improved Unconstrained Off-line Handwriting Recognition Gernot A. Fink, Thomas Plötz To cite this version: Gernot A. Fink, Thomas Plötz. Unsupervised

More information

A System for Searching and Browsing Spoken Communications

A System for Searching and Browsing Spoken Communications A System for Searching and Browsing Spoken Communications Lee Begeja Bernard Renger Murat Saraclar AT&T Labs Research 180 Park Ave Florham Park, NJ 07932 {lee, renger, murat} @research.att.com Abstract

More information

Real-Time Robot Audition System That Recognizes Simultaneous Speech in The Real World

Real-Time Robot Audition System That Recognizes Simultaneous Speech in The Real World Real-Time Robot Audition System That Recognizes Simultaneous Speech in The Real World Shun ichi Yamamoto, Kazuhiro Nakadai, Mikio Nakano, Hiroshi Tsujino, Jean-Marc Valin, Kazunori Komatani, Tetsuya Ogata,

More information

Portuguese Broadcast News and Statistical Machine Translation (SMT)

Portuguese Broadcast News and Statistical Machine Translation (SMT) Statistical Machine Translation of Broadcast News from Spanish to Portuguese Raquel Sánchez Martínez 1, João Paulo da Silva Neto 2, and Diamantino António Caseiro 1 L²F - Spoken Language Systems Laboratory,

More information

Training Ircam s Score Follower

Training Ircam s Score Follower Training Ircam s Follower Arshia Cont, Diemo Schwarz, Norbert Schnell To cite this version: Arshia Cont, Diemo Schwarz, Norbert Schnell. Training Ircam s Follower. IEEE International Conference on Acoustics,

More information

Slovak Automatic Dictation System for Judicial Domain

Slovak Automatic Dictation System for Judicial Domain Slovak Automatic Dictation System for Judicial Domain Milan Rusko 1(&), Jozef Juhár 2, Marián Trnka 1, Ján Staš 2, Sakhia Darjaa 1, Daniel Hládek 2, Róbert Sabo 1, Matúš Pleva 2, Marián Ritomský 1, and

More information

SENTIMENT EXTRACTION FROM NATURAL AUDIO STREAMS. Lakshmish Kaushik, Abhijeet Sangwan, John H. L. Hansen

SENTIMENT EXTRACTION FROM NATURAL AUDIO STREAMS. Lakshmish Kaushik, Abhijeet Sangwan, John H. L. Hansen SENTIMENT EXTRACTION FROM NATURAL AUDIO STREAMS Lakshmish Kaushik, Abhijeet Sangwan, John H. L. Hansen Center for Robust Speech Systems (CRSS), Eric Jonsson School of Engineering, The University of Texas

More information

have more skill and perform more complex

have more skill and perform more complex Speech Recognition Smartphone UI Speech Recognition Technology and Applications for Improving Terminal Functionality and Service Usability User interfaces that utilize voice input on compact devices such

More information

EFFICIENT HANDLING OF MULTILINGUAL LANGUAGE MODELS

EFFICIENT HANDLING OF MULTILINGUAL LANGUAGE MODELS EFFICIENT HANDLING OF MULTILINGUAL LANGUAGE MODELS Christian Fügen, Sebastian Stüker, Hagen Soltau, Florian Metze Interactive Systems Labs University of Karlsruhe Am Fasanengarten 5 76131 Karlsruhe, Germany

More information

Speech Transcription

Speech Transcription TC-STAR Final Review Meeting Luxembourg, 29 May 2007 Speech Transcription Jean-Luc Gauvain LIMSI TC-STAR Final Review Luxembourg, 29-31 May 2007 1 What Is Speech Recognition? Def: Automatic conversion

More information

Computer Assisted Language Learning (CALL) Systems

Computer Assisted Language Learning (CALL) Systems Tutorial on CALL in INTERSPEECH2012 by T.Kawahara and N.Minematsu 1 Computer Assisted Language g Learning (CALL) Systems Tatsuya Kawahara (Kyoto University, Japan) Nobuaki Minematsu (University of Tokyo,

More information

OPTIMIZATION OF NEURAL NETWORK LANGUAGE MODELS FOR KEYWORD SEARCH. Ankur Gandhe, Florian Metze, Alex Waibel, Ian Lane

OPTIMIZATION OF NEURAL NETWORK LANGUAGE MODELS FOR KEYWORD SEARCH. Ankur Gandhe, Florian Metze, Alex Waibel, Ian Lane OPTIMIZATION OF NEURAL NETWORK LANGUAGE MODELS FOR KEYWORD SEARCH Ankur Gandhe, Florian Metze, Alex Waibel, Ian Lane Carnegie Mellon University Language Technology Institute {ankurgan,fmetze,ahw,lane}@cs.cmu.edu

More information

Machine Learning with MATLAB David Willingham Application Engineer

Machine Learning with MATLAB David Willingham Application Engineer Machine Learning with MATLAB David Willingham Application Engineer 2014 The MathWorks, Inc. 1 Goals Overview of machine learning Machine learning models & techniques available in MATLAB Streamlining the

More information

Adaptive Training for Large Vocabulary Continuous Speech Recognition

Adaptive Training for Large Vocabulary Continuous Speech Recognition Adaptive Training for Large Vocabulary Continuous Speech Recognition Kai Yu Hughes Hall College and Cambridge University Engineering Department July 2006 Dissertation submitted to the University of Cambridge

More information

2014/02/13 Sphinx Lunch

2014/02/13 Sphinx Lunch 2014/02/13 Sphinx Lunch Best Student Paper Award @ 2013 IEEE Workshop on Automatic Speech Recognition and Understanding Dec. 9-12, 2013 Unsupervised Induction and Filling of Semantic Slot for Spoken Dialogue

More information

HMM-based Speech Synthesis with Various Degrees of Articulation: a Perceptual Study

HMM-based Speech Synthesis with Various Degrees of Articulation: a Perceptual Study HMM-based Speech Synthesis with Various Degrees of Articulation: a Perceptual Study Benjamin Picart, Thomas Drugman, Thierry Dutoit TCTS Lab, Faculté Polytechnique (FPMs), University of Mons (UMons), Belgium

More information

Automatic diacritization of Arabic for Acoustic Modeling in Speech Recognition

Automatic diacritization of Arabic for Acoustic Modeling in Speech Recognition Automatic diacritization of Arabic for Acoustic Modeling in Speech Recognition Dimitra Vergyri Speech Technology and Research Lab., SRI International, Menlo Park, CA 94025, USA dverg@speech.sri.com Katrin

More information

Leveraging Large Amounts of Loosely Transcribed Corporate Videos for Acoustic Model Training

Leveraging Large Amounts of Loosely Transcribed Corporate Videos for Acoustic Model Training Leveraging Large Amounts of Loosely Transcribed Corporate Videos for Acoustic Model Training Matthias Paulik and Panchi Panchapagesan Cisco Speech and Language Technology (C-SALT), Cisco Systems, Inc.

More information

Shogo Kumagai, Keisuke Doman, Tomokazu Takahashi, Daisuke Deguchi, Ichiro Ide, and Hiroshi Murase

Shogo Kumagai, Keisuke Doman, Tomokazu Takahashi, Daisuke Deguchi, Ichiro Ide, and Hiroshi Murase Detection of Inconsistency between Subject and Speaker based on the Co-occurrence of Lip Motion and Voice Towards Speech Scene Extraction from News Videos Shogo Kumagai, Keisuke Doman, Tomokazu Takahashi,

More information

Vector Quantization and Clustering

Vector Quantization and Clustering Vector Quantization and Clustering Introduction K-means clustering Clustering issues Hierarchical clustering Divisive (top-down) clustering Agglomerative (bottom-up) clustering Applications to speech recognition

More information

Weighting and Normalisation of Synchronous HMMs for Audio-Visual Speech Recognition

Weighting and Normalisation of Synchronous HMMs for Audio-Visual Speech Recognition ISCA Archive http://www.isca-speech.org/archive Auditory-Visual Speech Processing 27 (AVSP27) Hilvarenbeek, The Netherlands August 31 - September 3, 27 Weighting and Normalisation of Synchronous HMMs for

More information

Effect of Captioning Lecture Videos For Learning in Foreign Language 外 国 語 ( 英 語 ) 講 義 映 像 に 対 する 字 幕 提 示 の 理 解 度 効 果

Effect of Captioning Lecture Videos For Learning in Foreign Language 外 国 語 ( 英 語 ) 講 義 映 像 に 対 する 字 幕 提 示 の 理 解 度 効 果 Effect of Captioning Lecture Videos For Learning in Foreign Language VERI FERDIANSYAH 1 SEIICHI NAKAGAWA 1 Toyohashi University of Technology, Tenpaku-cho, Toyohashi 441-858 Japan E-mail: {veri, nakagawa}@slp.cs.tut.ac.jp

More information

MFCC-Based Voice Recognition System for Home Automation Using Dynamic Programming

MFCC-Based Voice Recognition System for Home Automation Using Dynamic Programming International Journal of Science and Research (IJSR) MFCC-Based Voice Recognition System for Home Automation Using Dynamic Programming Sandeep Joshi1, Sneha Nagar2 1 PG Student, Embedded Systems, Oriental

More information

Integration of Negative Emotion Detection into a VoIP Call Center System

Integration of Negative Emotion Detection into a VoIP Call Center System Integration of Negative Detection into a VoIP Call Center System Tsang-Long Pao, Chia-Feng Chang, and Ren-Chi Tsao Department of Computer Science and Engineering Tatung University, Taipei, Taiwan Abstract

More information

Using Adaptation to Improve Speech Transcription Alignment in Noisy and Reverberant Environments

Using Adaptation to Improve Speech Transcription Alignment in Noisy and Reverberant Environments Using Adaptation to Improve Speech Transcription Alignment in Noisy and Reverberant Environments Y. Mamiya 1, A. Stan 2, J. Yamagishi 1,3, P. Bell 1, O. Watts 1, R.A.J. Clark 1, S. King 1 1 Centre for

More information

EFFECTS OF TRANSCRIPTION ERRORS ON SUPERVISED LEARNING IN SPEECH RECOGNITION

EFFECTS OF TRANSCRIPTION ERRORS ON SUPERVISED LEARNING IN SPEECH RECOGNITION EFFECTS OF TRANSCRIPTION ERRORS ON SUPERVISED LEARNING IN SPEECH RECOGNITION By Ramasubramanian Sundaram A Thesis Submitted to the Faculty of Mississippi State University in Partial Fulfillment of the

More information

Strategies for Training Large Scale Neural Network Language Models

Strategies for Training Large Scale Neural Network Language Models Strategies for Training Large Scale Neural Network Language Models Tomáš Mikolov #1, Anoop Deoras 2, Daniel Povey 3, Lukáš Burget #4, Jan Honza Černocký #5 # Brno University of Technology, Speech@FIT,

More information

DynaSpeak: SRI s Scalable Speech Recognizer for Embedded and Mobile Systems

DynaSpeak: SRI s Scalable Speech Recognizer for Embedded and Mobile Systems DynaSpeak: SRI s Scalable Speech Recognizer for Embedded and Mobile Systems Horacio Franco, Jing Zheng, John Butzberger, Federico Cesari, Michael Frandsen, Jim Arnold, Venkata Ramana Rao Gadde, Andreas

More information

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics Proceedings of Meetings on Acoustics Volume 19, 2013 http://acousticalsociety.org/ ICA 2013 Montreal Montreal, Canada 2-7 June 2013 Speech Communication Session 2aSC: Linking Perception and Production

More information

Myanmar Continuous Speech Recognition System Based on DTW and HMM

Myanmar Continuous Speech Recognition System Based on DTW and HMM Myanmar Continuous Speech Recognition System Based on DTW and HMM Ingyin Khaing Department of Information and Technology University of Technology (Yatanarpon Cyber City),near Pyin Oo Lwin, Myanmar Abstract-

More information