ITERATIVE CONSTRAINED MLLR APPROACH FOR SPEAKER ADAPTATION

Size: px
Start display at page:

Download "ITERATIVE CONSTRAINED MLLR APPROACH FOR SPEAKER ADAPTATION"

Transcription

1 Proceedings of the IASTED International Conference Signal Processing, Pattern Recognition and Applications (SPPRA 3) February -, 3 Innsbruck, Austria ITERATIVE CONSTRAINED MLLR APPROACH FOR SPEAKER ADAPTATION Giorgio Biagetti, Alessandro Curzi, Massimo Mercuri, Claudio Turchetti DII Dipartimento di Ingegneria dell Informazione, Università Politecnica delle Marche, I-3 Ancona, Italy. g.biagetti@univpm.it, a.curzi@univpm.it, massimo.mercuri@univpm.it, c.turchetti@univpm.it ABSTRACT In this paper an effective technique for speaker adaptation on the feature domain is presented. This technique starts from the well known maximum-likelihood linear regression (MLLR) auxiliary function to obtain the constrained MLLR transformation in an iterative fashion. The proposed approach is particularly suitable to be implemented on the client side of a distributed speech recognition scheme, due to the reduced number of iterations required to reach convergence. Extensive experimentation using the CMU Sphinx ASR system along with a preliminarily trained speaker-independent acoustic model for the Italian language, in a setting designed for large-vocabulary continuous speech recognition, demonstrates the effectiveness of the approach even with small amounts of adaptation data. KEY WORDS MLLR,,, DSR, SI, SD. Introduction Speaker adaptation techniques have proven to be very effective in modern speech recognition systems [], especially when there are significant mismatches between the training and decoding conditions. In these techniques one starts with a speaker-independent (SI) model, and then tries to accommodate the model to a new speaker to obtain a speaker-dependent (SD) model, using a relatively small amount of speech data from the new speaker. The basic idea is to compensate for the mismatch between training and test conditions by modifying the model parameters on the basis of some adaptation data. Among these techniques the maximum-likelihood linear regression (MLLR) [] and constrained MLLR () [9],[] are powerful and widely used methods for speaker adaptation in large-vocabulary continuous speech recognition (LVCSR). MLLR uses the expectation-maximization (EM) criterion to estimate a linear transformation to adapt Gaussian parameters, i.e. their mean and variance, of hidden Markov models (HMMs) []. Although the two transformations are estimated separately, the computational complexity is reasonably high. An alternative scheme to adapt both mean vectors and covariance matrices is to use a approach, in which the transformation applied to the covariance matrix corresponds to the transformation applied to the mean vector. It can be shown that is equivalent to a transformation in the feature domain. This property makes CM- LLR particularly suitable in a distributed speech recognition (DSR) scheme, in which the recognition process is split up into a front-end on the client side primarily related to feature extraction, and a back-end on the server side devoted to the recognition itself. The main drawbacks of the approach are: the algorithm is more complex than MLLR; it is an iterative process which converges usually after about 3 iterations, but in some cases it does not converge even after iterations [3]. Thus, as the algorithm for the implementation of CM- LLR is more complex than standard MLLR, there is a need for simpler algorithms to be efficiently implemented on the client side of a DSR scheme. The proposed algorithm meets this requirement, and due to simpler formulation is able to overcome some of the limitations of the. It s worth noting that the iterative has exactly the same formulation as the MLLR algorithm, while requiring less iterations than to converge. The algorithm has been evaluated by extensive experimentation using the CMU Sphinx recognizer in a setting defined for LVCSR and performance comparison with MLLR and techniques shows the effectiveness of the approach. MLLR and background Both MLLR and use the EM criterion to estimate a linear transformation to adapt the Gaussian parameters of HMMs. Starting from the current set of parameters M, the adapted model parameters ˆM are obtained by maximizing the following auxiliary function: Q(M, ˆM) = K γ m (τ)[k m +log( ˆΣ m ) + (o(τ) ˆµ m ) T ˆΣ m (o(τ) ˆµ m )], () where ˆµ m and ˆΣ m are the adapted mean and variance of component m for the target acoustic condition while M and DOI:.3/P

2 T represent respectively the number of components associated with the particular transform and the number of observations. K is a constant dependent only on the transition probabilities, K m is the normalisation constant associated with Gaussian component m, and γ m (τ) = p(q m (τ) M, O T ) () is the posterior occupancy of component m, being q m (τ) the Gaussian m at time τ and O T = [o(),..., o(t )] the observation sequence.. Unconstrained transformation In this adaptation method the mean and variance are transformed independently of each other. The mean µ is transformed as: ˆµ = Aµ + b = W ξ, (3) where ξ is the extended mean vector, [ µ T ] T, and W = b A] is the extended linear transform. The transform of the covariance matrix Σ is given by: ˆΣ = HΣH T, () where H is the matrix to be estimated. Equation () represents the objective function to be maximized during adaptation to obtain the parameters W and H of the transformations. It was ly proposed to adapt the mean vector [], extending the technique to variance adaptation only later []. The mean based linear transform is referred to as MLLR, while covariance matrix transform is named variance MLLR.. Constrained transformation The mean and the variance MLLR transformations can be simultaneously applied to both mean vectors and covariance matrices. However, as in this case the computational cost is high, a constrained scheme to adapt both mean vectors and covariance matrices can be used [, 7]. This is referred to as constrained MLLR: ˆµ = Aµ + b, () ˆΣ = AΣA T, () which is a particular case of unconstrained transformation with H = A. By substituting () and () into equation () and assuming a diagonal covariance matrix Σ, the following auxiliary function to be maximized is obtained: Q(M, ˆM) = K where γ m (τ)[k m +log( Σ m ) log( A ) + (ô(τ) µ m ) T Σ m (ô(τ) µ m )], (7) ô(τ) = A o(τ) A b = A c o(τ)+b c = W c ζ(τ) () as usual W c = [b c A c ] represents the extended matrix transformation, and ζ(τ) = [ o(τ) T ] T is the extended vector of observations. Equation (7) clearly shows that the constrained transformation can be directly applied in the feature domain []. 3 Iterative The proposed algorithm, referred to as iterative (), is able to implement the constrained transformation using the standard auxiliary function () for MLLR, instead of maximizing the more complex objective function (7). The transform estimation is an iterative process: a first transformation W is estimated by MLLR given an initial estimate of Σ, then at each iteration a new estimation ˆΣ k is forced to be ˆΣ k = A k ˆΣk A T k (9) until convergence is reached. The algorithm proceeds as follows:. Assume an initial estimate Σ of ˆΣ.. Estimate the mean transformation W k = [b k A k ] by equation (). 3. A new estimate is obtained by the constraint (9).. If a stop criterion on both ˆµ and ˆΣ is not met, return to step.. Otherwise, if a stop criterion on both ˆµ and ˆΣ is met, the solution is reached and the transformation () on feature domain can be applied. Fig. shows a block diagram of the algorithm. An example of the effect of its operation can be found in Figs. and 7. In particular, Fig. shows as a reference the effect of applying a conventional algorithm to the 3 MFCCs of one short utterance chosen among those used in our experiments. The graphs also report an aligned phonetic transcription of the selected utterance, the dark area marking a short silence period within. Fig. 7 reports analogous results for our algorithm. As can be seen, the conventional adaptation algorithm appears not to be able to substantially adapt the first few (three) components, while the proposed algorithm is able not only to adapt these components, but also to displace the adapted observation vector more. We will see in section that recognition performance with the proposed adaptation is higher, which led us to believe that the larger displacement was able to move the cepstral vector towards those of the baseline SI model more effectively. Computational cost The complexity of an iteration is of the same order as the complexity of MLLR. Given an estimate of Σ, 397

3 Assume an initial estimate Σ of ˆΣ Estimate the mean transform W k = [b k A k ] Stopping criterion met? YES Apply transformation on feature domain ô(τ) = W c ζ(τ) NO ˆΣ k = A k ˆΣk A T k Figure : Block diagram of the algorithm. at each iteration the computes the transformation W, which can be obtained by solving γ m (τ)σ m o(τ)ξ T m = = γ m (τ)σ m W ξ m ξ T m () For the full covariance matrix case the solution is computationally very expensive, however, for the diagonal case a closed-form solution is computationally feasible []. The left-hand side of equation () is independent of the transformation matrix and will be referred to as Z, where: Z = A new variable G i is defined as G i = m= and W is calculated using γ m (τ)σ m o(τ)ξ T m () σm ξ m ξm T i w i T γ m (τ) () τ= = G T i z i (3) where w i is the i-th row of W and z i is the i-th row of Z. Solving equation (3) requires the inversion of an (n+) (n+) matrix for each row of W, being n the size of the mean vectors. As the matrix inversion takes O(n 3 ) operations, thus the estimation of W requires O(n ) operations at each iteration. Once the transformation is obtained, O(M n) operations are required to achieve a new estimation of ˆΣ by transformation (9). Thus the total computational cost for each iteration is approximately given by O(n ) + O(Mn) O(n ). n corresponds also to the number of Mel-frequency cepstral coefficients (MFCC) plus and, which is how the input vectors are compounded. Table : Parameters used in the experiments. baseline acoustic model: training language: Italian total audio length: 9 hours states per HMM: 3 + final state Gaussians per state: tied states: test corpus: language: Italian book title: I promessi sposi, by Alessandro Manzoni book chapter: source: Liber Liber ( total number of utterances: average number of phones per utterance: front-end: standard: ETSI ES audio: ks/s, mono, bit features: 3 MFCC + + language model: 3-gram statistical model In, optimising the auxiliary function (7) with respect to W leads to the update formulae. It has been shown in [9] that the i-th row of W is given by: w i = (αp i + k i )G i () where p i is the cofactor [ c i... c in ], (i.e. c ij = cof(a ij )) and G i = k i = σ m= m i τ= m= σ m i µ mi Given the total occupancy β = γ m (τ)ζ(τ)ζ(τ) T, () γ m (τ)ζ(τ) T. () τ= γ m (τ), (7) the coefficient α satisfies the following quadratic expression: α p T i G i p i + αp T i G i k i β =. () This is a simple quadratic expression in α and may be solved in the usual way. The main cost in estimating W is due to the computation of cofactors. Every row requires 3 n3 + n + n operations, thus the computational cost is of the order of O(n ) per iteration. This neglects the actual cost of inverting G i which only needs to be performed once, costing O(n ). Unfortunately the constrained case use an indirect optimisation scheme. The total cost then becomes (I + )O(n ) where I is the total number of iterations. In reality of course when using incremental adaptation the new transform estimate is initialised with the previous one, thus dramatically reducing the required number of iterations. Furthermore, it is not necessary to invert G i, as an indirect optimisation over each row may be used. 39

4 Iterative Conventional MLLR 3 Iterative WER [%] Iterative relative error 3 word error rate [%] relative error [%] word error rate [%] 3 number of iterations Figure : Accuracy evaluation as a function of the number of iterations. Adaptation was carried out with the first test corpus utterance, while the remaining material was used for recognition purposes. 3 number of iterations Figure 3: relative error and recognition accuracy evaluation as a function of the number of iterations. Test setup is the same as for Fig. Experimental results In order to verify the effectiveness of the algorithm, experiments were conducted using the CMU Sphinx ASR system, together with an advanced ETSI ES feature extractor. The setup used in the experiments is reported in Tab.. The SI baseline model was generated according to the method described in []. All the experiments reported in this section were conducted using the first chapter of a long audiobook in Italian, whose audio and text transcriptions are freely available. In the first experiment, the iteration process was initialized by adapting the model on the first utterance of the test corpus, while the remaining utterances had been left available for recognition purposes. Each utterance was expected to be phones long, corresponding to an average duration in time of about ETSI frames (. s). After initialization, the transform W c = [b c A c ] was estimated according the iteration process depicted in Fig.. Then the matrix W c was used to transform the features as shown in equation (). The estimation accuracy of ICM- LLR was evaluated by performing several recognition tests, and comparing the results with those obtained by MLLR and conventional. The behavior reported in Fig. shows that the word error rate (WER), defined as the ratio of wrongly recognized or missing words to total words in the text, reduces for both constrained algorithms as the number of iterations increases, slightly approaching the reference MLLR accuracy. It also must be noted that behaves always better than. In addition, as a means to evaluate the convergence, the behavior of the relative error as a function of the number of iterations can be derived as well. The relative error for the iterative estimation of matrix W at step k is defined as e k % = W k W k W W (9) and can be used when defining a stopping criterion. Fig. 3 shows that the relative error decreases remarkably as the number of iterations increases, following the error rate trend. A second experiment was performed to assess whether the recognition accuracy improved using an increasing number of utterances as adaptation data. All the results were compared with those obtained with conventional by setting the number of iterations for both algorithms at. As can be seen from Fig., the approach shows better performance in terms of word error rate, and thus gives a better adaptation. The algorithm was also tested in an incremental online adaptation framework, collecting data during all the recognition task. This setup is commonly used when the transcription is not available like in spoken dialog applications. Thus, as soon as a new incoming utterance becomes available, a new transformation is applied to the previously adapted model. As with any other online MLLR adaptation approach, and as confirmed by the results in Fig., in this case was not stable and the model needed to be restored periodically, requiring an objective comparison with the baseline [3]. Nevertheless, even if only for a limited number of updates, the accuracy kept improving. Conclusion This work presents an algorithm performing speaker adaptation on the feature domain, particularly suitable to be employed on the client side of a DSR scheme. Starting from the general MLLR auxiliary function, the proposed tech- 399

5 Iterative Conventional 3 3 Iterative Model restoration points word error rate[%] word error rate [%] 3 number of utterances Figure : Accuracy comparison between and. The number of iterations was fixed at for both algorithms while letting adaptation data increase one utterance at a time. incoming utterance number Figure : accuracy performance in an incremental adaptation framework. Model was updated each time a new utterance became available. The large square dots depict the WER when a restore operation was performed. nique is able to implement the constrained transformation on an iterative basis. Several adaptation tests on a preliminarily trained SI baseline model lead to a notable recognition performance already after very few iterations, thus demonstrating the effectiveness of the approach. Comparisons with the widely used conventional constrained MLLR, showed that ICM- LLR improves convergence rate, and thus the overall computational complexity, yet maintaining a slight benefit in terms of word error rate even with a small amount of adaptation data. The proposed method was also tested on an on-line adaptation scenario, giving promising though quite preliminary results. Further work need to be done in this area in order to improve the long-term stability of the system. References [] D. Povey and K. Yao, A basis method for robust estimation of constrained MLLR, Proc. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP-), Prague, Czech Republic, May, 3. [] L. R. Rabiner, A tutorial on hidden Markov models and selected applications in speech recognition, Proc. IEEE, 77(), 99, 7. [3] Y. Li, H. Erdogan, Y. Gao, and E. Marcheret, Incremental on-line feature space MLLR adaptation for telephony speech recognition, Proc. 7th International Conference on Spoken Language Processing (ICSLP - Interspeech ), Denver, Colorado,, 7. density hidden Markov models, Computer Speech and Language, 9(), 99, 7. [] M. Gales and P. Woodland, Mean and variance adaptation within the MLLR framework, Computer Speech and Language, (), 99, 9. [] V. Digalakis, D. Rtischev, and L. Neumeyer, Speaker adaptation using constrained estimation of Gaussian mixtures, IEEE Transactions on Speech and Audio Processing, 3(), 99, [7] M. Ferras, C. C. Leung, C. Barras, and J.-L. Gauvain, Constrained MLLR for speaker recognition, in IEEE International Conference on Acoustics, Speech and Signal Processing,7. ICASSP 7.,, 7, IV-3 IV-. [] M. Ferras, C.-C. Leung, C. Barras, and J.-L. Gauvain, Comparison of speaker adaptation methods as feature extraction for SVM-based speaker recognition, IEEE Transactions on Audio, Speech, and Language Processing, (),, [9] M. Gales, Maximum likelihood linear transformations for HMM-based speech recognition, Computer Speech and Language, (), 99, 7 9. [] M. Alessandrini, G. Biagetti, A. Curzi, C. Turchetti, Semi-automatic acoustic model generation from large unsynchronized audio and text chunks, Proc. th Annual Conference of the International Speech Communication Association (Interspeech ), Florence, Italy,,. [] C. Leggetter and P. Woodland, Maximum likelihood linear regression for speaker adaptation of continuous

6 k w e l r a m o d e l l a ě o d i k o m o [s]k e v o l à e a m e ţ o à o r n o k w e l r a m o d e l l a ě o d i k o m o [s]k e v o l à e a m e ţ o à o r n o MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # Figure : Comparison of the MFCC features and the same features after adaptation with the conventional algorithm. Above the graphs is shown the aligned phonetic transcription of the utterance under consideration. Figure 7: Comparison of the MFCC features and the same features after adaptation with the proposed ICM- LLR algorithm. Above the graphs is shown the aligned phonetic transcription of the utterance under consideration.

7 MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # MFCC # Figure : continued. Figure 7: continued.

Membering T M : A Conference Call Service with Speaker-Independent Name Dialing on AIN

Membering T M : A Conference Call Service with Speaker-Independent Name Dialing on AIN PAGE 30 Membering T M : A Conference Call Service with Speaker-Independent Name Dialing on AIN Sung-Joon Park, Kyung-Ae Jang, Jae-In Kim, Myoung-Wan Koo, Chu-Shik Jhon Service Development Laboratory, KT,

More information

AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS

AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS PIERRE LANCHANTIN, ANDREW C. MORRIS, XAVIER RODET, CHRISTOPHE VEAUX Very high quality text-to-speech synthesis can be achieved by unit selection

More information

Hardware Implementation of Probabilistic State Machine for Word Recognition

Hardware Implementation of Probabilistic State Machine for Word Recognition IJECT Vo l. 4, Is s u e Sp l - 5, Ju l y - Se p t 2013 ISSN : 2230-7109 (Online) ISSN : 2230-9543 (Print) Hardware Implementation of Probabilistic State Machine for Word Recognition 1 Soorya Asokan, 2

More information

SOME ASPECTS OF ASR TRANSCRIPTION BASED UNSUPERVISED SPEAKER ADAPTATION FOR HMM SPEECH SYNTHESIS

SOME ASPECTS OF ASR TRANSCRIPTION BASED UNSUPERVISED SPEAKER ADAPTATION FOR HMM SPEECH SYNTHESIS SOME ASPECTS OF ASR TRANSCRIPTION BASED UNSUPERVISED SPEAKER ADAPTATION FOR HMM SPEECH SYNTHESIS Bálint Tóth, Tibor Fegyó, Géza Németh Department of Telecommunications and Media Informatics Budapest University

More information

Developing an Isolated Word Recognition System in MATLAB

Developing an Isolated Word Recognition System in MATLAB MATLAB Digest Developing an Isolated Word Recognition System in MATLAB By Daryl Ning Speech-recognition technology is embedded in voice-activated routing systems at customer call centres, voice dialling

More information

Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition

Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition , Lisbon Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition Wolfgang Macherey Lars Haferkamp Ralf Schlüter Hermann Ney Human Language Technology

More information

Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics

Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics Anastasis Kounoudes 1, Anixi Antonakoudi 1, Vasilis Kekatos 2 1 The Philips College, Computing and Information Systems

More information

Speech Recognition on Cell Broadband Engine UCRL-PRES-223890

Speech Recognition on Cell Broadband Engine UCRL-PRES-223890 Speech Recognition on Cell Broadband Engine UCRL-PRES-223890 Yang Liu, Holger Jones, John Johnson, Sheila Vaidya (Lawrence Livermore National Laboratory) Michael Perrone, Borivoj Tydlitat, Ashwini Nanda

More information

Speech: A Challenge to Digital Signal Processing Technology for Human-to-Computer Interaction

Speech: A Challenge to Digital Signal Processing Technology for Human-to-Computer Interaction : A Challenge to Digital Signal Processing Technology for Human-to-Computer Interaction Urmila Shrawankar Dept. of Information Technology Govt. Polytechnic, Nagpur Institute Sadar, Nagpur 440001 (INDIA)

More information

SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA

SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA Nitesh Kumar Chaudhary 1 and Shraddha Srivastav 2 1 Department of Electronics & Communication Engineering, LNMIIT, Jaipur, India 2 Bharti School Of Telecommunication,

More information

IEEE Proof. Web Version. PROGRESSIVE speaker adaptation has been considered

IEEE Proof. Web Version. PROGRESSIVE speaker adaptation has been considered IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING 1 A Joint Factor Analysis Approach to Progressive Model Adaptation in Text-Independent Speaker Verification Shou-Chun Yin, Richard Rose, Senior

More information

Establishing the Uniqueness of the Human Voice for Security Applications

Establishing the Uniqueness of the Human Voice for Security Applications Proceedings of Student/Faculty Research Day, CSIS, Pace University, May 7th, 2004 Establishing the Uniqueness of the Human Voice for Security Applications Naresh P. Trilok, Sung-Hyuk Cha, and Charles C.

More information

Using Adaptation to Improve Speech Transcription Alignment in Noisy and Reverberant Environments

Using Adaptation to Improve Speech Transcription Alignment in Noisy and Reverberant Environments Using Adaptation to Improve Speech Transcription Alignment in Noisy and Reverberant Environments Y. Mamiya 1, A. Stan 2, J. Yamagishi 1,3, P. Bell 1, O. Watts 1, R.A.J. Clark 1, S. King 1 1 Centre for

More information

EVALUATION OF AUTOMATIC TRANSCRIPTION SYSTEMS FOR THE JUDICIAL DOMAIN

EVALUATION OF AUTOMATIC TRANSCRIPTION SYSTEMS FOR THE JUDICIAL DOMAIN EVALUATION OF AUTOMATIC TRANSCRIPTION SYSTEMS FOR THE JUDICIAL DOMAIN J. Lööf (1), D. Falavigna (2),R.Schlüter (1), D. Giuliani (2), R. Gretter (2),H.Ney (1) (1) Computer Science Department, RWTH Aachen

More information

7 Gaussian Elimination and LU Factorization

7 Gaussian Elimination and LU Factorization 7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method

More information

ADVANCES IN ARABIC BROADCAST NEWS TRANSCRIPTION AT RWTH. David Rybach, Stefan Hahn, Christian Gollan, Ralf Schlüter, Hermann Ney

ADVANCES IN ARABIC BROADCAST NEWS TRANSCRIPTION AT RWTH. David Rybach, Stefan Hahn, Christian Gollan, Ralf Schlüter, Hermann Ney ADVANCES IN ARABIC BROADCAST NEWS TRANSCRIPTION AT RWTH David Rybach, Stefan Hahn, Christian Gollan, Ralf Schlüter, Hermann Ney Human Language Technology and Pattern Recognition Computer Science Department,

More information

Ericsson T18s Voice Dialing Simulator

Ericsson T18s Voice Dialing Simulator Ericsson T18s Voice Dialing Simulator Mauricio Aracena Kovacevic, Anna Dehlbom, Jakob Ekeberg, Guillaume Gariazzo, Eric Lästh and Vanessa Troncoso Dept. of Signals Sensors and Systems Royal Institute of

More information

ADAPTIVE AND DISCRIMINATIVE MODELING FOR IMPROVED MISPRONUNCIATION DETECTION. Horacio Franco, Luciana Ferrer, and Harry Bratt

ADAPTIVE AND DISCRIMINATIVE MODELING FOR IMPROVED MISPRONUNCIATION DETECTION. Horacio Franco, Luciana Ferrer, and Harry Bratt ADAPTIVE AND DISCRIMINATIVE MODELING FOR IMPROVED MISPRONUNCIATION DETECTION Horacio Franco, Luciana Ferrer, and Harry Bratt Speech Technology and Research Laboratory, SRI International, Menlo Park, CA

More information

Turkish Radiology Dictation System

Turkish Radiology Dictation System Turkish Radiology Dictation System Ebru Arısoy, Levent M. Arslan Boaziçi University, Electrical and Electronic Engineering Department, 34342, Bebek, stanbul, Turkey arisoyeb@boun.edu.tr, arslanle@boun.edu.tr

More information

BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION

BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION P. Vanroose Katholieke Universiteit Leuven, div. ESAT/PSI Kasteelpark Arenberg 10, B 3001 Heverlee, Belgium Peter.Vanroose@esat.kuleuven.ac.be

More information

Online Diarization of Telephone Conversations

Online Diarization of Telephone Conversations Odyssey 2 The Speaker and Language Recognition Workshop 28 June July 2, Brno, Czech Republic Online Diarization of Telephone Conversations Oshry Ben-Harush, Itshak Lapidot, Hugo Guterman Department of

More information

Advanced Signal Processing and Digital Noise Reduction

Advanced Signal Processing and Digital Noise Reduction Advanced Signal Processing and Digital Noise Reduction Saeed V. Vaseghi Queen's University of Belfast UK WILEY HTEUBNER A Partnership between John Wiley & Sons and B. G. Teubner Publishers Chichester New

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

More information

Comp 14112 Fundamentals of Artificial Intelligence Lecture notes, 2015-16 Speech recognition

Comp 14112 Fundamentals of Artificial Intelligence Lecture notes, 2015-16 Speech recognition Comp 14112 Fundamentals of Artificial Intelligence Lecture notes, 2015-16 Speech recognition Tim Morris School of Computer Science, University of Manchester 1 Introduction to speech recognition 1.1 The

More information

By choosing to view this document, you agree to all provisions of the copyright laws protecting it.

By choosing to view this document, you agree to all provisions of the copyright laws protecting it. This material is posted here with permission of the IEEE Such permission of the IEEE does not in any way imply IEEE endorsement of any of Helsinki University of Technology's products or services Internal

More information

Trading Strategies and the Cat Tournament Protocol

Trading Strategies and the Cat Tournament Protocol M A C H I N E L E A R N I N G P R O J E C T F I N A L R E P O R T F A L L 2 7 C S 6 8 9 CLASSIFICATION OF TRADING STRATEGIES IN ADAPTIVE MARKETS MARK GRUMAN MANJUNATH NARAYANA Abstract In the CAT Tournament,

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Gaussian Mixture Models Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique

More information

Statistical Machine Translation: IBM Models 1 and 2

Statistical Machine Translation: IBM Models 1 and 2 Statistical Machine Translation: IBM Models 1 and 2 Michael Collins 1 Introduction The next few lectures of the course will be focused on machine translation, and in particular on statistical machine translation

More information

How to Improve the Sound Quality of Your Microphone

How to Improve the Sound Quality of Your Microphone An Extension to the Sammon Mapping for the Robust Visualization of Speaker Dependencies Andreas Maier, Julian Exner, Stefan Steidl, Anton Batliner, Tino Haderlein, and Elmar Nöth Universität Erlangen-Nürnberg,

More information

Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus

Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives

More information

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression Logistic Regression Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max

More information

Automatic Transcription of Conversational Telephone Speech

Automatic Transcription of Conversational Telephone Speech IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 13, NO. 6, NOVEMBER 2005 1173 Automatic Transcription of Conversational Telephone Speech Thomas Hain, Member, IEEE, Philip C. Woodland, Member, IEEE,

More information

Lecture 3: Finding integer solutions to systems of linear equations

Lecture 3: Finding integer solutions to systems of linear equations Lecture 3: Finding integer solutions to systems of linear equations Algorithmic Number Theory (Fall 2014) Rutgers University Swastik Kopparty Scribe: Abhishek Bhrushundi 1 Overview The goal of this lecture

More information

Video Affective Content Recognition Based on Genetic Algorithm Combined HMM

Video Affective Content Recognition Based on Genetic Algorithm Combined HMM Video Affective Content Recognition Based on Genetic Algorithm Combined HMM Kai Sun and Junqing Yu Computer College of Science & Technology, Huazhong University of Science & Technology, Wuhan 430074, China

More information

General Framework for an Iterative Solution of Ax b. Jacobi s Method

General Framework for an Iterative Solution of Ax b. Jacobi s Method 2.6 Iterative Solutions of Linear Systems 143 2.6 Iterative Solutions of Linear Systems Consistent linear systems in real life are solved in one of two ways: by direct calculation (using a matrix factorization,

More information

1 Solving LPs: The Simplex Algorithm of George Dantzig

1 Solving LPs: The Simplex Algorithm of George Dantzig Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.

More information

TED-LIUM: an Automatic Speech Recognition dedicated corpus

TED-LIUM: an Automatic Speech Recognition dedicated corpus TED-LIUM: an Automatic Speech Recognition dedicated corpus Anthony Rousseau, Paul Deléglise, Yannick Estève Laboratoire Informatique de l Université du Maine (LIUM) University of Le Mans, France firstname.lastname@lium.univ-lemans.fr

More information

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

More information

Question 2: How do you solve a matrix equation using the matrix inverse?

Question 2: How do you solve a matrix equation using the matrix inverse? Question : How do you solve a matrix equation using the matrix inverse? In the previous question, we wrote systems of equations as a matrix equation AX B. In this format, the matrix A contains the coefficients

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical

More information

Objective Intelligibility Assessment of Text-to-Speech Systems Through Utterance Verification

Objective Intelligibility Assessment of Text-to-Speech Systems Through Utterance Verification Objective Intelligibility Assessment of Text-to-Speech Systems Through Utterance Verification Raphael Ullmann 1,2, Ramya Rasipuram 1, Mathew Magimai.-Doss 1, and Hervé Bourlard 1,2 1 Idiap Research Institute,

More information

How To Recognize Voice Over Ip On Pc Or Mac Or Ip On A Pc Or Ip (Ip) On A Microsoft Computer Or Ip Computer On A Mac Or Mac (Ip Or Ip) On An Ip Computer Or Mac Computer On An Mp3

How To Recognize Voice Over Ip On Pc Or Mac Or Ip On A Pc Or Ip (Ip) On A Microsoft Computer Or Ip Computer On A Mac Or Mac (Ip Or Ip) On An Ip Computer Or Mac Computer On An Mp3 Recognizing Voice Over IP: A Robust Front-End for Speech Recognition on the World Wide Web. By C.Moreno, A. Antolin and F.Diaz-de-Maria. Summary By Maheshwar Jayaraman 1 1. Introduction Voice Over IP is

More information

HMM-based Speech Synthesis with Various Degrees of Articulation: a Perceptual Study

HMM-based Speech Synthesis with Various Degrees of Articulation: a Perceptual Study HMM-based Speech Synthesis with Various Degrees of Articulation: a Perceptual Study Benjamin Picart, Thomas Drugman, Thierry Dutoit TCTS Lab, Faculté Polytechnique (FPMs), University of Mons (UMons), Belgium

More information

Linear Programming. March 14, 2014

Linear Programming. March 14, 2014 Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1

More information

Adaptive Training for Large Vocabulary Continuous Speech Recognition

Adaptive Training for Large Vocabulary Continuous Speech Recognition Adaptive Training for Large Vocabulary Continuous Speech Recognition Kai Yu Hughes Hall College and Cambridge University Engineering Department July 2006 Dissertation submitted to the University of Cambridge

More information

LEARNING FEATURE MAPPING USING DEEP NEURAL NETWORK BOTTLENECK FEATURES FOR DISTANT LARGE VOCABULARY SPEECH RECOGNITION

LEARNING FEATURE MAPPING USING DEEP NEURAL NETWORK BOTTLENECK FEATURES FOR DISTANT LARGE VOCABULARY SPEECH RECOGNITION LEARNING FEATURE MAPPING USING DEEP NEURAL NETWORK BOTTLENECK FEATURES FOR DISTANT LARGE VOCABULARY SPEECH RECOGNITION Ivan Himawan 1, Petr Motlicek 1, David Imseng 1, Blaise Potard 1, Namhoon Kim 2, Jaewon

More information

1 Determinants and the Solvability of Linear Systems

1 Determinants and the Solvability of Linear Systems 1 Determinants and the Solvability of Linear Systems In the last section we learned how to use Gaussian elimination to solve linear systems of n equations in n unknowns The section completely side-stepped

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

Roots of Equations (Chapters 5 and 6)

Roots of Equations (Chapters 5 and 6) Roots of Equations (Chapters 5 and 6) Problem: given f() = 0, find. In general, f() can be any function. For some forms of f(), analytical solutions are available. However, for other functions, we have

More information

Predict the Popularity of YouTube Videos Using Early View Data

Predict the Popularity of YouTube Videos Using Early View Data 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Subspace Analysis and Optimization for AAM Based Face Alignment

Subspace Analysis and Optimization for AAM Based Face Alignment Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China zhaoming1999@zju.edu.cn Stan Z. Li Microsoft

More information

Analysis of an Artificial Hormone System (Extended abstract)

Analysis of an Artificial Hormone System (Extended abstract) c 2013. This is the author s version of the work. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purpose or for creating

More information

3.2. Solving quadratic equations. Introduction. Prerequisites. Learning Outcomes. Learning Style

3.2. Solving quadratic equations. Introduction. Prerequisites. Learning Outcomes. Learning Style Solving quadratic equations 3.2 Introduction A quadratic equation is one which can be written in the form ax 2 + bx + c = 0 where a, b and c are numbers and x is the unknown whose value(s) we wish to find.

More information

Practical Guide to the Simplex Method of Linear Programming

Practical Guide to the Simplex Method of Linear Programming Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear

More information

CS 688 Pattern Recognition Lecture 4. Linear Models for Classification

CS 688 Pattern Recognition Lecture 4. Linear Models for Classification CS 688 Pattern Recognition Lecture 4 Linear Models for Classification Probabilistic generative models Probabilistic discriminative models 1 Generative Approach ( x ) p C k p( C k ) Ck p ( ) ( x Ck ) p(

More information

STATISTICA Formula Guide: Logistic Regression. Table of Contents

STATISTICA Formula Guide: Logistic Regression. Table of Contents : Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary

More information

Speech Recognition System of Arabic Alphabet Based on a Telephony Arabic Corpus

Speech Recognition System of Arabic Alphabet Based on a Telephony Arabic Corpus Speech Recognition System of Arabic Alphabet Based on a Telephony Arabic Corpus Yousef Ajami Alotaibi 1, Mansour Alghamdi 2, and Fahad Alotaiby 3 1 Computer Engineering Department, King Saud University,

More information

Available from Deakin Research Online:

Available from Deakin Research Online: This is the authors final peered reviewed (post print) version of the item published as: Adibi,S 2014, A low overhead scaled equalized harmonic-based voice authentication system, Telematics and informatics,

More information

DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson

DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson c 1999 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or

More information

Background: State Estimation

Background: State Estimation State Estimation Cyber Security of the Smart Grid Dr. Deepa Kundur Background: State Estimation University of Toronto Dr. Deepa Kundur (University of Toronto) Cyber Security of the Smart Grid 1 / 81 Dr.

More information

Factor Analysis. Chapter 420. Introduction

Factor Analysis. Chapter 420. Introduction Chapter 420 Introduction (FA) is an exploratory technique applied to a set of observed variables that seeks to find underlying factors (subsets of variables) from which the observed variables were generated.

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes

More information

Automatic parameter regulation for a tracking system with an auto-critical function

Automatic parameter regulation for a tracking system with an auto-critical function Automatic parameter regulation for a tracking system with an auto-critical function Daniela Hall INRIA Rhône-Alpes, St. Ismier, France Email: Daniela.Hall@inrialpes.fr Abstract In this article we propose

More information

Linear Threshold Units

Linear Threshold Units Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear

More information

10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method

10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method 578 CHAPTER 1 NUMERICAL METHODS 1. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS As a numerical technique, Gaussian elimination is rather unusual because it is direct. That is, a solution is obtained after

More information

Automatic slide assignation for language model adaptation

Automatic slide assignation for language model adaptation Automatic slide assignation for language model adaptation Applications of Computational Linguistics Adrià Agustí Martínez Villaronga May 23, 2013 1 Introduction Online multimedia repositories are rapidly

More information

Robust Methods for Automatic Transcription and Alignment of Speech Signals

Robust Methods for Automatic Transcription and Alignment of Speech Signals Robust Methods for Automatic Transcription and Alignment of Speech Signals Leif Grönqvist (lgr@msi.vxu.se) Course in Speech Recognition January 2. 2004 Contents Contents 1 1 Introduction 2 2 Background

More information

Nonlinear Iterative Partial Least Squares Method

Nonlinear Iterative Partial Least Squares Method Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for

More information

A Robustness Simulation Method of Project Schedule based on the Monte Carlo Method

A Robustness Simulation Method of Project Schedule based on the Monte Carlo Method Send Orders for Reprints to reprints@benthamscience.ae 254 The Open Cybernetics & Systemics Journal, 2014, 8, 254-258 Open Access A Robustness Simulation Method of Project Schedule based on the Monte Carlo

More information

Linear Equations ! 25 30 35$ & " 350 150% & " 11,750 12,750 13,750% MATHEMATICS LEARNING SERVICE Centre for Learning and Professional Development

Linear Equations ! 25 30 35$ &  350 150% &  11,750 12,750 13,750% MATHEMATICS LEARNING SERVICE Centre for Learning and Professional Development MathsTrack (NOTE Feb 2013: This is the old version of MathsTrack. New books will be created during 2013 and 2014) Topic 4 Module 9 Introduction Systems of to Matrices Linear Equations Income = Tickets!

More information

Sensitivity Analysis 3.1 AN EXAMPLE FOR ANALYSIS

Sensitivity Analysis 3.1 AN EXAMPLE FOR ANALYSIS Sensitivity Analysis 3 We have already been introduced to sensitivity analysis in Chapter via the geometry of a simple example. We saw that the values of the decision variables and those of the slack and

More information

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc. 1. Introduction Linear Programming for Optimization Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc. 1.1 Definition Linear programming is the name of a branch of applied mathematics that

More information

MISSING FEATURE RECONSTRUCTION AND ACOUSTIC MODEL ADAPTATION COMBINED FOR LARGE VOCABULARY CONTINUOUS SPEECH RECOGNITION

MISSING FEATURE RECONSTRUCTION AND ACOUSTIC MODEL ADAPTATION COMBINED FOR LARGE VOCABULARY CONTINUOUS SPEECH RECOGNITION MISSING FEATURE RECONSTRUCTION AND ACOUSTIC MODEL ADAPTATION COMBINED FOR LARGE VOCABULARY CONTINUOUS SPEECH RECOGNITION Ulpu Remes, Kalle J. Palomäki, and Mikko Kurimo Adaptive Informatics Research Centre,

More information

Solving Mass Balances using Matrix Algebra

Solving Mass Balances using Matrix Algebra Page: 1 Alex Doll, P.Eng, Alex G Doll Consulting Ltd. http://www.agdconsulting.ca Abstract Matrix Algebra, also known as linear algebra, is well suited to solving material balance problems encountered

More information

Emotion Detection from Speech

Emotion Detection from Speech Emotion Detection from Speech 1. Introduction Although emotion detection from speech is a relatively new field of research, it has many potential applications. In human-computer or human-human interaction

More information

OPTIMIZATION OF NEURAL NETWORK LANGUAGE MODELS FOR KEYWORD SEARCH. Ankur Gandhe, Florian Metze, Alex Waibel, Ian Lane

OPTIMIZATION OF NEURAL NETWORK LANGUAGE MODELS FOR KEYWORD SEARCH. Ankur Gandhe, Florian Metze, Alex Waibel, Ian Lane OPTIMIZATION OF NEURAL NETWORK LANGUAGE MODELS FOR KEYWORD SEARCH Ankur Gandhe, Florian Metze, Alex Waibel, Ian Lane Carnegie Mellon University Language Technology Institute {ankurgan,fmetze,ahw,lane}@cs.cmu.edu

More information

Grant: LIFE08 NAT/GR/000539 Total Budget: 1,664,282.00 Life+ Contribution: 830,641.00 Year of Finance: 2008 Duration: 01 FEB 2010 to 30 JUN 2013

Grant: LIFE08 NAT/GR/000539 Total Budget: 1,664,282.00 Life+ Contribution: 830,641.00 Year of Finance: 2008 Duration: 01 FEB 2010 to 30 JUN 2013 Coordinating Beneficiary: UOP Associated Beneficiaries: TEIC Project Coordinator: Nikos Fakotakis, Professor Wire Communications Laboratory University of Patras, Rion-Patras 26500, Greece Email: fakotaki@upatras.gr

More information

Myanmar Continuous Speech Recognition System Based on DTW and HMM

Myanmar Continuous Speech Recognition System Based on DTW and HMM Myanmar Continuous Speech Recognition System Based on DTW and HMM Ingyin Khaing Department of Information and Technology University of Technology (Yatanarpon Cyber City),near Pyin Oo Lwin, Myanmar Abstract-

More information

Solving Systems of Linear Equations Using Matrices

Solving Systems of Linear Equations Using Matrices Solving Systems of Linear Equations Using Matrices What is a Matrix? A matrix is a compact grid or array of numbers. It can be created from a system of equations and used to solve the system of equations.

More information

Automatic Detection of Laughter and Fillers in Spontaneous Mobile Phone Conversations

Automatic Detection of Laughter and Fillers in Spontaneous Mobile Phone Conversations Automatic Detection of Laughter and Fillers in Spontaneous Mobile Phone Conversations Hugues Salamin, Anna Polychroniou and Alessandro Vinciarelli University of Glasgow - School of computing Science, G128QQ

More information

KL-DIVERGENCE REGULARIZED DEEP NEURAL NETWORK ADAPTATION FOR IMPROVED LARGE VOCABULARY SPEECH RECOGNITION

KL-DIVERGENCE REGULARIZED DEEP NEURAL NETWORK ADAPTATION FOR IMPROVED LARGE VOCABULARY SPEECH RECOGNITION KL-DIVERGENCE REGULARIZED DEEP NEURAL NETWORK ADAPTATION FOR IMPROVED LARGE VOCABULARY SPEECH RECOGNITION Dong Yu 1, Kaisheng Yao 2, Hang Su 3,4, Gang Li 3, Frank Seide 3 1 Microsoft Research, Redmond,

More information

Experiments with Signal-Driven Symbolic Prosody for Statistical Parametric Speech Synthesis

Experiments with Signal-Driven Symbolic Prosody for Statistical Parametric Speech Synthesis Experiments with Signal-Driven Symbolic Prosody for Statistical Parametric Speech Synthesis Fabio Tesser, Giacomo Sommavilla, Giulio Paci, Piero Cosi Institute of Cognitive Sciences and Technologies, National

More information

The Filtered-x LMS Algorithm

The Filtered-x LMS Algorithm The Filtered-x LMS Algorithm L. Håkansson Department of Telecommunications and Signal Processing, University of Karlskrona/Ronneby 372 25 Ronneby Sweden Adaptive filters are normally defined for problems

More information

Regression III: Advanced Methods

Regression III: Advanced Methods Lecture 16: Generalized Additive Models Regression III: Advanced Methods Bill Jacoby Michigan State University http://polisci.msu.edu/jacoby/icpsr/regress3 Goals of the Lecture Introduce Additive Models

More information

HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER

HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER Gholamreza Anbarjafari icv Group, IMS Lab, Institute of Technology, University of Tartu, Tartu 50411, Estonia sjafari@ut.ee

More information

2014/02/13 Sphinx Lunch

2014/02/13 Sphinx Lunch 2014/02/13 Sphinx Lunch Best Student Paper Award @ 2013 IEEE Workshop on Automatic Speech Recognition and Understanding Dec. 9-12, 2013 Unsupervised Induction and Filling of Semantic Slot for Spoken Dialogue

More information

Session 7 Bivariate Data and Analysis

Session 7 Bivariate Data and Analysis Session 7 Bivariate Data and Analysis Key Terms for This Session Previously Introduced mean standard deviation New in This Session association bivariate analysis contingency table co-variation least squares

More information

2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system

2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system 1. Systems of linear equations We are interested in the solutions to systems of linear equations. A linear equation is of the form 3x 5y + 2z + w = 3. The key thing is that we don t multiply the variables

More information

Politecnico di Torino. Porto Institutional Repository

Politecnico di Torino. Porto Institutional Repository Politecnico di Torino Porto Institutional Repository [Proceeding] Multiport network analyzer self-calibration: a new approach and some interesting results Original Citation: G.L. Madonna, A. Ferrero, U.

More information

Artificial Neural Network for Speech Recognition

Artificial Neural Network for Speech Recognition Artificial Neural Network for Speech Recognition Austin Marshall March 3, 2005 2nd Annual Student Research Showcase Overview Presenting an Artificial Neural Network to recognize and classify speech Spoken

More information

Airport Planning and Design. Excel Solver

Airport Planning and Design. Excel Solver Airport Planning and Design Excel Solver Dr. Antonio A. Trani Professor of Civil and Environmental Engineering Virginia Polytechnic Institute and State University Blacksburg, Virginia Spring 2012 1 of

More information

Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices

Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices Matrices 2. Solving Square Systems of Linear Equations; Inverse Matrices Solving square systems of linear equations; inverse matrices. Linear algebra is essentially about solving systems of linear equations,

More information

Generating Training Data for Medical Dictations

Generating Training Data for Medical Dictations Generating Training Data for Medical Dictations Sergey Pakhomov University of Minnesota, MN pakhomov.sergey@mayo.edu Michael Schonwetter Linguistech Consortium, NJ MSchonwetter@qwest.net Joan Bachenko

More information

METHODOLOGICAL CONSIDERATIONS OF DRIVE SYSTEM SIMULATION, WHEN COUPLING FINITE ELEMENT MACHINE MODELS WITH THE CIRCUIT SIMULATOR MODELS OF CONVERTERS.

METHODOLOGICAL CONSIDERATIONS OF DRIVE SYSTEM SIMULATION, WHEN COUPLING FINITE ELEMENT MACHINE MODELS WITH THE CIRCUIT SIMULATOR MODELS OF CONVERTERS. SEDM 24 June 16th - 18th, CPRI (Italy) METHODOLOGICL CONSIDERTIONS OF DRIVE SYSTEM SIMULTION, WHEN COUPLING FINITE ELEMENT MCHINE MODELS WITH THE CIRCUIT SIMULTOR MODELS OF CONVERTERS. Áron Szûcs BB Electrical

More information

DERIVATIVES AS MATRICES; CHAIN RULE

DERIVATIVES AS MATRICES; CHAIN RULE DERIVATIVES AS MATRICES; CHAIN RULE 1. Derivatives of Real-valued Functions Let s first consider functions f : R 2 R. Recall that if the partial derivatives of f exist at the point (x 0, y 0 ), then we

More information

Linearly Independent Sets and Linearly Dependent Sets

Linearly Independent Sets and Linearly Dependent Sets These notes closely follow the presentation of the material given in David C. Lay s textbook Linear Algebra and its Applications (3rd edition). These notes are intended primarily for in-class presentation

More information

Lean Six Sigma Analyze Phase Introduction. TECH 50800 QUALITY and PRODUCTIVITY in INDUSTRY and TECHNOLOGY

Lean Six Sigma Analyze Phase Introduction. TECH 50800 QUALITY and PRODUCTIVITY in INDUSTRY and TECHNOLOGY TECH 50800 QUALITY and PRODUCTIVITY in INDUSTRY and TECHNOLOGY Before we begin: Turn on the sound on your computer. There is audio to accompany this presentation. Audio will accompany most of the online

More information

Problem of Missing Data

Problem of Missing Data VASA Mission of VA Statisticians Association (VASA) Promote & disseminate statistical methodological research relevant to VA studies; Facilitate communication & collaboration among VA-affiliated statisticians;

More information

Exploratory Factor Analysis

Exploratory Factor Analysis Introduction Principal components: explain many variables using few new variables. Not many assumptions attached. Exploratory Factor Analysis Exploratory factor analysis: similar idea, but based on model.

More information

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its

More information