Creation of a Speech to Text System for Kiswahili
|
|
|
- Felix Lawson
- 10 years ago
- Views:
Transcription
1 Creation of a Speech to Text System for Kiswahili Dr. Katherine Getao and Evans Miriti University of Nairobi Abstract The creation of a speech to text system for any language is an onerous task. This is especially the case if not much has been done in the area of speech processing for that language previously. The lack of or little availability of previous work means that there are few resources on which one can build upon. A speech to text system can be considered to consist of four main components. These are the front end, the emission model, lexicon and finally, a component that combines these three to perform transcription from acoustic signal to text. In this paper we will describe the implementation of a Kiswahili lexicon, emission model, and the training for these components. We also describe the implementation of the Viterbi algorithm that we chose to use for our transcription. We will also describe any limitations of our approach and give suggestions for improvement. Introduction Dictation can be described as the transcription of extended monologue by a single speaker. Dictation systems fall within the realm of large vocabulary continuous speech recognition systems (LVCSR), which means they have a vocabulary of over 5000 words (Jurafsky and Martin, 2000).There are several speech dictation systems now available in the market. Some of these systems can provide transcription rates of as much as 160 words per minute, word accuracy about 99 % and vocabularies of up to 250,000 words ( Two of the most commonly used Dictation systems in the market are and IBM s ViaVoice Pro USB edition with 98.4% word accuracy and Dragon Naturally speaking V 8.0 with 98.6% word accuracy, ( ). These commercial systems are limited in the languages that they support. For instance ViaVoice Pro supports German, Italian, Japanese, UK English and US English ( There is little evidence of existence of dictation software for Kiswahili. Our aim therefore is to take pioneering steps towards the development of a robust and accurate dictation system for Kiswahili. In this paper, we describe three main components of a simple dictation system we built to transcribe Kiswahili speech to text. The system recognizes speech only by a single speaker but it
2 is our intention to expand it further to recognize speech by any speaker. The system is sentence based, which means that the input to the system will be an acoustic signal representing a single sentence. The system is sentence based which means that the input to the system will be an acoustic signal representing a single sentence. We are going to describe how we went about implementing the different components of the system. We also give some of the results obtained and finally give some suggestions on how the system can be improved. Components of a Speech to Text System To build a speech to text system, one has to build four main components. These are the: 1. Front end The main function of the front end is to pre-process the speech signal and extract speech features from the each segment of the acoustic signal. These features are the input to the emission model. For the input features, we use Mel-frequency Cepstral Coefficients (ETSI, 2003). 2. Emission Model The purpose of this component is to find the correspondence between the input speech features and a given phone. It seeks to find out the probability that a given segment (represented by the speech features) represents a given phone 3. Lexicon The lexicon contains all the words in the language. Any sentence that is produced will be formed using some words in the lexicon in a given sequence 4. The transcriber The component puts all the other components together to produce the final sentence. We are going to describe how we built the emission model, the lexicon and the transcriber. We will describe the challenges we faced and how we went about overcoming them. For each component, we will state its shortcomings and how they can be overcome to improve its performance. Emission Model Our emission model should produce the probability P(o v i ) Eq 1 which is the probability of observation o given input v i. o is a vector of mel-frequency cepestral coefficients (MFCC) features for a given segment of the acoustic signal and v i is one of the phones in the language. The emission model generates this probability for each phone v i i=1 N where N is the number of phones in the language. For Kiswahili, N=31 (Mohammed, 2001). For our emission model we have added 3 more symbols to use as phones in our language. These are si for silence, x a sound not found in Kiswahili but which is necessary for the transcription of foreign words in the language. And gh which occurs in our training data. Thus our N=34. We use multi-layer perceptron (MLP) neural network to create the emission model. A MLP is trained using the back-propagation algorithm. We use examples of spoken sentences as training data. Each example consists of:
3 1. A wave file representing the acoustic signal generated when a given sentence is spoken 2. A text file labeled with the correct transcription (phonetic symbols) for each time segment of the acoustic signal. We have to process this data to a format that can be used to train the neural network. For each segment of the wave file, we generate the MFCC features. This is the input to the neural network. We also generate the probability for each phone given the segment. Since we know the correct phone for the segment for the labels file, we give this phone the highest probability. In our case, we give the phone used to label the segment a probability of 0.8 and all the other phones share a probability of 0.2 i.e. 0.2/33 for each phone. This is the expected output of the neural network. Each example in our training data contains several hundreds of segments. So we generate the inputs and expected output for each segment in each file for all the files. This is the training data for our neural network. We use this data to train the neural network. After training, our neural network should be able to generate the correct probability for phones, given a vector of speech features for a previously unseen segment if it has learnt to generalize correctly. The phone with the highest probability will be assumed to be the correct label for that segment. The probability generated by the neural network is P(v i o) Eq 2 i.e. probability of phone v i given input o. The probability we require is P(o v i ). This can be obtained using Bayes theorem. The formula for obtaining the required probability is given below. P( vi o) P( o) P ( o vi ) = P( v ) i P(o) is constant for all phones. Thus it can be done away with since it amount to multiply all P( vi o) the probabilities we have with a constant factor. This gives us the equation P ( o vi ) =. P( v ) We thus need P(v i o) and P(v i ). P(v i o) is the probability generated by the neural network. P(v i ) is obtained by dividing the number of occurrences of phone v i in the training data by the total number of phones in the training data (Holmes, 2001). Lexicon The lexicon contains the words that the transcribed words will be made of. These words are obtained from a corpus. Our lexicon needs to represent the following parameters: 1. The words in the language in their transcribed format 2. The bigram probabilities between the words 3. The probability of each word starting a sentence Representing the words Kiswahili is made up of 31 phonemes and each of these phonemes corresponds directly to a written form. For instance the sound /ʃ/ always corresponds to the orthographic form sh as in i
4 the word shule (school). The sound /ʧ/ always corresponds to the written form ch as in the word chatu (python). It s for this reason that Kiswahili phonemes can be represented by their orthographic form rather than use a phonetic alphabet. Using this format, the orthographic form of a word can be assumed to be its phonetically transcribed format. The word shule for instance consists of four phonemes i.e. /sh/, /u/, /l/ and /e/. Using this representation, each phone can be represented by a number from 0 to 30. Thus in our lexicon, each word is represented by the sequence of numbers representing the phones that make up the phonemes of the word. To get the word in orthographic form, we just get the orthographic form represented by each number and combine all the forms. This representation saves space because the actual words don t have to be represented. It also helps in that numbers are easier to work with for instance in comparison than strings. In addition to the sequence of words representing the phonemes of the word, we also have a unique number for each word. This is achieved by ordering the words we have alphabetically in ascending order then giving each number a unique id starting from zero. This is important in particular when we form bigram probability connections between the words. We just need to say the a word with id i is connected to a word with a id j by a bigram probability p. Bigram Probabilities Representation As stated above, we need to represent the bigram probabilities between the words as obtained from a corpus then smoothed. Ideally, if the corpus doesn t suffer from sparseness, each word i in the corpus should be connected by a bigram probability b ij to every other word j in the corpus. But sparseness is a problem that besets every corpus and it s dealt with using smoothing techniques. Given a word word i and every other word word j in the corpus with which it occurs in the sequence word i word j in the corpus i.e. word j after word i, then we have a bigram probability between the words that we have to represent in our lexicon. This is achieved by having two arrays in the representation of word i. The first array contains the ids of the words that word i has a bigram probability with. The second array contains the actual bigram probability between word i and word j in the same position as the id of word j in the first array. This is elaborated using the figure below.
5 j 1 j 2 j 3 j n b ij1 b ij2 b ij3 b ijn Figure 1: Representation of bigram probabilities between word with id i and words with ids j 1, j 2,,j n. We will also need a default probability between word with id i and any other word with id k with which it doesn t occur in sequence with. This is a single value for all these words represented by a variable of type double. The probability of each word starting a sentence This again is a simple value obtained from the corpus. By getting the number of times a word starts a sentence in the corpus and dividing by the number of sentences. Some words don t start any sentence in the corpus. Smoothing techniques are used to give these words a small probability of starting a sentence. Lexicon Word Entry The entry of each word in the lexicon is thus an object made of five components 1. An integer representing the id of the word 2. An array of short integers representing the phones making up the word 3. A variable of type double representing the probability of the word starting a sentence 4. Two arrays of the form shown in figure 1 showing the bigram probabilities between the word and other words with which it occurs in sequence with in the corpus. 5. A variable of type double representing the probability of the word forming a bigram (in which it is the first word) with any other word with which it doesn t occur in sequence with in the corpus. The lexicon is thus an array of the objects described above ordered by the word ids. The Recognizer The recognizer is the component that uses all the other components and the probabilities they provide to come up with the sequence of words making up a sentence. We implemented the recognizing function using the Viterbi algorithm. Details of the viterbi algorithm can be found in Holmes (2001) and Jurafsky and Martin (2000). Here a summary of our actual implementation is given. Each word in the lexicon contains a number of phones (say N w ). Each of these phones corresponds to a state identified by the word id and the position of the phone in the word. The Viterbi algorithm starts by creating a matrix where the number of rows is equal to the total
6 number of states (phones) for all the words in the lexicon and the number of columns is equal to the total number of inputs (analogous to number of segments) in the audio signal. Since we need to cater for the non-emitting initial state and final state (since they are not produced by any inputs) the number of columns can be increased by two. We can refer to the total number of states as N and the total number of columns as M. in our implementation, we just have two columns. We also make the assumption that each word will start from the first phone rather than from any of its phones. Remember that due to deletions, its possible for the first phone of a word not to be uttered. This possibility is ignored. The two columns are represented as arrays of states. Each state is an object referred to as viterbi node, each viterbi node contains a probability which is the likely hood that we are in that state given that we are in the m th input. Each state also has an array of states (a path) we have passed through to get to that state. Initially (when we haven t seen any input) Each state (in column 1) corresponding to the first phone of a word is given a probability which is equal to the probability of the word starting a sentence obtained from the lexicon. This is the probability that the sentence we will get starts with that word. Note that column 1 and column 2 contain identical states although the probabilities for column 2 are set to zero. When we get the first input, we need to move to column 2. We use the procedure below to move to column 2: For each input k in the vectors of inputs For each state i in column 1 whose probability is not zero: For each state j in column 2 Get the probability that the current input was produced by the phone corresponding to this state (j) (use the emission model) Get the transition probability from state i (in column 1) to state j (in column 2) Get the probability in state i (in column 1) Multiply the three probabilities to get the new probability if this new probability is greater than the current probability in node j (in column 2), then this probability becomes the current probability of node j. We also indicate that the current probability was obtained by a transition from node i. //end for each state j //end for each state i For each state j in column 2 update the path of the states we have passed through to get to this state by adding the node i from column 1 which we got our current probability from.
7 Interchange the columns, column 1 becomes column 2 and column 2 becomes column 1. Set the probabilities of column 2 to zero //end for each input k // this part is executed after the for each k loop. We have gone through all the inputs. Get the node in column 2 with the highest probability. (this is the highest probability path) Get the path in this node of the nodes we passed through. (Since each entry in the path corresponds to a phone and a word id, we get the sequence of word ids. This gives us the list of words we passed through. This is our recognized sentence.) The transitions that any state can have are described below i. Transition to itself (loop back) ii. A transition to a state coming after it in the word iii. If the state is a at the end of a word, then it can transit either to itself or to the first state in another word Limiting the number of transitions reduces the computation time. The first and the second probabilities are obtained by counting the number of times a phone follows itself in the labeled speech files (c1)and the number of times it is followed by other phones (c2) c1/ c1+ c2 gives the first probability c2/ c1+ c2 gives the second probability The transition probability from the last phone to the first phone in another word is the bigram probability between the words
8 Results and Discussion Table 1: Examples of transcribed sentences Correct Sentence Eee bwana weee si unafahamu tena mambo ya presha Yeye ni kati ya waliofukuzwa alidai kuwa amepania kukata rufaa mbele ya kamati ya shule hiyo kwa madai kwamba alikamatwa akiingia kwenye bweni la wanawake Usiku baada ya kuzidiwa na kilevi Baada ya kudaiwa kuhusika na kifo hicho mzee huyo alibanwa na waganga watatu wa jadi Misi huyo aliyekuwa mrembo wa kwanza kuiwakilisha Tanzania Kona ya mahaba Endelea kupata raha ya huba sehemu hii Haya katika twisheni yetu inayoendelea leo kwa juu kidogo kisha tembea nayo kwenda mashariki na magharibi Si kuna upenyo mkubwa pale ndugu yangu hebu inase ile sehemu ya nyuma ya kisigino Kidole chako kikubwa cha mguuni na kile kinachofuata Transcribed Sentence ebu wao tei tu si unafahamu tena mambo ya presha vile ni kati ya waliofukuzwa alidai kuwa amepania kukata rufaa mbele li kamati lilo kwa madai kwamba alikamatwa akiingia kwenye bweni la wanawake usiku baada ya kuzidiwa na kile baba kwa leo kuhusika na kifo hicho mzee huyo uliokita alibanwa na waganga watatu wa bali Misi huyo aliyekuwa mrembo wa kwanza kuiwakilisha Tanzania kona ambao ni kati ya kukata kubwa zile ambao ni kamwe ya katika twisheni tu bali maoni ya lilo kona kwa juu kidogo kisha tembea nayo kwenda mashariki na magharibi kwa tu siku na roba upenyo mkubwa tu bali bilashi tei mambo binafsi si miongoni mwa kisigino kidale tei chako kikubwa cha mguuni na kile kinachofuata
9 Figure 2: The variation of the validation error with the number of training examples. Each example is in the form of a labeled speech file Figure 2 shows that the validation error for the emission model goes down as the number of training examples increases. The emission model uses a Multi Layer Perceptron learning component. The general sloping of the graph to the right is an indication that improvement in performance of the application can be obtained by increasing the number of labeled speech files used to train the emission model. Figure 3: Variation of word error rate with the number of training examples.
10 Figure 3 shows that the word error rate goes down as the number of training examples. Thus we need to acquire more training data in the form of labeled speech files Figure 4: Variation of word error rate with the validation error. Figure 4 shows a general downward trend in the word error rate with a reduction in the validation error. The smaller error rates can be obtained by using more training examples. More results need to be obtained to find if the trend should continue with lower validation error or the graph will take on a general rising trend at some point. The rugged nature of the curve can be corrected by use of averages. Achievements 1. Creation of a system that successfully transcribes Kiswahili sentences to text for a single speaker. The best word error rate obtained is 43%. 2. Creation of a lexicon of about 800 words for the Kiswahili ASR system. These are the only words that can appear in sentences produced by the Kiswahili Dictation system. The method used to construct this lexicon can be reused for construction of bigger lexicons both for Kiswahili and for other languages. reports that dragon naturally speaking has a vocabulary of about 250,000 words. 3. Creation of a speech processing front end that analysis an audio signal to produce melfrequency spectral coefficients (MFCC). The class for perfoming MFCC analysis takes one speech segment at a time and produces the corresponding speech features in an array. Thus this class can be used for other speech processing applications e.g. speaker identification 4. Creation of a class to calculate both the levenshtein distance between strings and the word error rate.
11 Limitations 1. High word error rate 43%. Though most of the sentences produced are intelligible, and some contain no errors at all, some contain too many errors thus driving the word error rate high, thus more work needs to be done to reduce the word error rate 2. Low speed. The transcription speed is quite low. This is the reason the vocabulary (number of words in the lexicon had to be kept low). Faster transcription methods need to be used to improve the speed to acceptable rates. A useful transcription system should be able to transcribe at a higher speed than the average user can type. 3. Small vocabulary. The vocabulary had to be kept small due to the slow transcription speed. The structure of the lexicon allows the vocabulary to grow to any size. 4. Slow training rate The training speed for the emission model neural network is too slow. This may call for either a change in the learning element 5. Number of speakers The system was trained to recognize only the voice of one speaker. There is need to train the system with data obtained from many speakers so that it can perform well when tested with speech from any Kiswahili speaker. Conclusion The main objective of building a simple Kiswahili dictation system was accomplished. In the process, a lot of the techniques for building ASR systems were learnt. It is our aim to keep working on this in order to overcome its major shortcomings. In addition, we expect that the components we produce will make it easier for us to develop ASR applications for other African languages. Further Work 1. Creation of a labeled speech corpus for Kiswahili Standard speech corpuses are important in ASR research because they can be used for training and testing of the system. It then becomes possible to compare results from different systems using the same data set (Gold and Morgan, 2004). It would thus be useful to create a relatively large labeled speech corpus containing speech data from different speakers that can serve this role for Kiswahili. An example of a standard corpus for English is TIMIT. It contains data obtained by recording 630 speakers with each speaker recorded for 10 sentences one of which is spoken twice. 2. Use of biphones and or triphones The basic speech unit in this system is the phone. The disadvantage of using of using phones is that they fail to model co-articulation effects. To overcome this, either biphones or triphones can be used. In the case of biphones, the speech units are obtained by getting all the possible
12 permutations of phones in the language taking two at a time. For Kiswahili (with 31 phones), this would be about 961 units. The large number of units can be reduced by taking into consideration the fact that some combinations don t occur in the language. For triphones, all the possible permutations of the phones in the language taking three at a time is used. 3. Use of faster transcription algorithm For this system, the Viterbi algorithm is used for transcription. Another algorithm that can be used is the A* (A star) algorithm. Since this is a best first search algorithm, it should give a faster transcription speed. 4. Increase the size of the vocabulary The size of the vocabulary can be made larger once the speed of transcription is improved. 5. Comparison of performance of different learning elements Other types of learning elements e.g. support vector machines can be used and their performance compared with the performance of a neural network based learning element. References 1. Croft, A. Davison, R. Hargreaves, M. (2001).Engineering Mathematics. A Foundation For Electronic, Electrical, Communications and Systems Engineers. Pearson Education Limited. Harlow Essex, England. 2. Dyslexic.com. (2006). Comparison of Speech Recognition Software. [Online]. Available: 3. Gakuru, M. Iraki, F.K. Tucker, R. Shalanova K. Ngugi, K. Design of Speech Data Base for Unit Selection in Kiswahili TTS. [Online]. Available: 4. ETSI. (2003), Speech Processing, Transmission and Quality Aspects (STQ); Distributed speech recognition;front-end feature extraction algorithm;compression algorithms. [Online] Available: 5. Gold, B. And Morgan, N. (2004). Speech and Audio signal Processing. John Wiley And Sons, Singapore. 6. Holmes, J. And Holmes, W. (2001). Speech Synthesis and Recognition. Taylor And Francis.London. 7. Holmes, W. And Huckvale, M. (1994). Why Have HMMS Been So Successful For Automatic Speech Recognition And How Might They Be Improved? [Online]. Available: 8. Jurafsky, D. And Martin J. (2000). Speech And Language Processing. Pearson Education, Singapore. 9. Mitchell, M.M. (1997). Machine Learning. McGraw-Hill, Singapore. 10. Mohammed, M.A. (2001). Modern Swahili Grammar. East African Educational Publishers. Nairobi. 11. Rabiner Lawrence. (1989). A Tutorial On Hidden Markov Models And Selected Applications In Speech Recognition. Proceedings Of The IEEE, 77(2), SpeechMagic. [Online] Available:
13 13. Stephen Cook. (2001) Speech Recognition HOWTO. [Online]. Available: The Engineering Toolbox. (2005). Acoustics, Acoustic properties, Decibel ABC, Noise Rating (NR) curves, Sound transmission, Attenuation. [Online] Available: The Kamusi Project Internet Living Swahili Dictionary. [Online] Available: The CMU Sphinx Group Open Source Speech Recognition Engines. [Online] Available:
Turkish Radiology Dictation System
Turkish Radiology Dictation System Ebru Arısoy, Levent M. Arslan Boaziçi University, Electrical and Electronic Engineering Department, 34342, Bebek, stanbul, Turkey [email protected], [email protected]
AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS
AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS PIERRE LANCHANTIN, ANDREW C. MORRIS, XAVIER RODET, CHRISTOPHE VEAUX Very high quality text-to-speech synthesis can be achieved by unit selection
Speech: A Challenge to Digital Signal Processing Technology for Human-to-Computer Interaction
: A Challenge to Digital Signal Processing Technology for Human-to-Computer Interaction Urmila Shrawankar Dept. of Information Technology Govt. Polytechnic, Nagpur Institute Sadar, Nagpur 440001 (INDIA)
Artificial Neural Network for Speech Recognition
Artificial Neural Network for Speech Recognition Austin Marshall March 3, 2005 2nd Annual Student Research Showcase Overview Presenting an Artificial Neural Network to recognize and classify speech Spoken
Spot me if you can: Uncovering spoken phrases in encrypted VoIP conversations
Spot me if you can: Uncovering spoken phrases in encrypted VoIP conversations C. Wright, L. Ballard, S. Coull, F. Monrose, G. Masson Talk held by Goran Doychev Selected Topics in Information Security and
Lecture 12: An Overview of Speech Recognition
Lecture : An Overview of peech Recognition. Introduction We can classify speech recognition tasks and systems along a set of dimensions that produce various tradeoffs in applicability and robustness. Isolated
Establishing the Uniqueness of the Human Voice for Security Applications
Proceedings of Student/Faculty Research Day, CSIS, Pace University, May 7th, 2004 Establishing the Uniqueness of the Human Voice for Security Applications Naresh P. Trilok, Sung-Hyuk Cha, and Charles C.
Comp 14112 Fundamentals of Artificial Intelligence Lecture notes, 2015-16 Speech recognition
Comp 14112 Fundamentals of Artificial Intelligence Lecture notes, 2015-16 Speech recognition Tim Morris School of Computer Science, University of Manchester 1 Introduction to speech recognition 1.1 The
An Arabic Text-To-Speech System Based on Artificial Neural Networks
Journal of Computer Science 5 (3): 207-213, 2009 ISSN 1549-3636 2009 Science Publications An Arabic Text-To-Speech System Based on Artificial Neural Networks Ghadeer Al-Said and Moussa Abdallah Department
CONATION: English Command Input/Output System for Computers
CONATION: English Command Input/Output System for Computers Kamlesh Sharma* and Dr. T. V. Prasad** * Research Scholar, ** Professor & Head Dept. of Comp. Sc. & Engg., Lingaya s University, Faridabad, India
Speech Recognition System of Arabic Alphabet Based on a Telephony Arabic Corpus
Speech Recognition System of Arabic Alphabet Based on a Telephony Arabic Corpus Yousef Ajami Alotaibi 1, Mansour Alghamdi 2, and Fahad Alotaiby 3 1 Computer Engineering Department, King Saud University,
Myanmar Continuous Speech Recognition System Based on DTW and HMM
Myanmar Continuous Speech Recognition System Based on DTW and HMM Ingyin Khaing Department of Information and Technology University of Technology (Yatanarpon Cyber City),near Pyin Oo Lwin, Myanmar Abstract-
Estonian Large Vocabulary Speech Recognition System for Radiology
Estonian Large Vocabulary Speech Recognition System for Radiology Tanel Alumäe, Einar Meister Institute of Cybernetics Tallinn University of Technology, Estonia October 8, 2010 Alumäe, Meister (TUT, Estonia)
Hardware Implementation of Probabilistic State Machine for Word Recognition
IJECT Vo l. 4, Is s u e Sp l - 5, Ju l y - Se p t 2013 ISSN : 2230-7109 (Online) ISSN : 2230-9543 (Print) Hardware Implementation of Probabilistic State Machine for Word Recognition 1 Soorya Asokan, 2
NLP Programming Tutorial 5 - Part of Speech Tagging with Hidden Markov Models
NLP Programming Tutorial 5 - Part of Speech Tagging with Hidden Markov Models Graham Neubig Nara Institute of Science and Technology (NAIST) 1 Part of Speech (POS) Tagging Given a sentence X, predict its
Ericsson T18s Voice Dialing Simulator
Ericsson T18s Voice Dialing Simulator Mauricio Aracena Kovacevic, Anna Dehlbom, Jakob Ekeberg, Guillaume Gariazzo, Eric Lästh and Vanessa Troncoso Dept. of Signals Sensors and Systems Royal Institute of
Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition
, Lisbon Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition Wolfgang Macherey Lars Haferkamp Ralf Schlüter Hermann Ney Human Language Technology
SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS
UDC: 004.8 Original scientific paper SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS Tonimir Kišasondi, Alen Lovren i University of Zagreb, Faculty of Organization and Informatics,
Data quality in Accounting Information Systems
Data quality in Accounting Information Systems Comparing Several Data Mining Techniques Erjon Zoto Department of Statistics and Applied Informatics Faculty of Economy, University of Tirana Tirana, Albania
Robust Methods for Automatic Transcription and Alignment of Speech Signals
Robust Methods for Automatic Transcription and Alignment of Speech Signals Leif Grönqvist ([email protected]) Course in Speech Recognition January 2. 2004 Contents Contents 1 1 Introduction 2 2 Background
Back Propagation Neural Networks User Manual
Back Propagation Neural Networks User Manual Author: Lukáš Civín Library: BP_network.dll Runnable class: NeuralNetStart Document: Back Propagation Neural Networks Page 1/28 Content: 1 INTRODUCTION TO BACK-PROPAGATION
Objective Intelligibility Assessment of Text-to-Speech Systems Through Utterance Verification
Objective Intelligibility Assessment of Text-to-Speech Systems Through Utterance Verification Raphael Ullmann 1,2, Ramya Rasipuram 1, Mathew Magimai.-Doss 1, and Hervé Bourlard 1,2 1 Idiap Research Institute,
Open-Source, Cross-Platform Java Tools Working Together on a Dialogue System
Open-Source, Cross-Platform Java Tools Working Together on a Dialogue System Oana NICOLAE Faculty of Mathematics and Computer Science, Department of Computer Science, University of Craiova, Romania [email protected]
C E D A T 8 5. Innovating services and technologies for speech content management
C E D A T 8 5 Innovating services and technologies for speech content management Company profile 25 years experience in the market of transcription/reporting services; Cedat 85 Group: Cedat 85 srl Subtitle
Efficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words
, pp.290-295 http://dx.doi.org/10.14257/astl.2015.111.55 Efficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words Irfan
Membering T M : A Conference Call Service with Speaker-Independent Name Dialing on AIN
PAGE 30 Membering T M : A Conference Call Service with Speaker-Independent Name Dialing on AIN Sung-Joon Park, Kyung-Ae Jang, Jae-In Kim, Myoung-Wan Koo, Chu-Shik Jhon Service Development Laboratory, KT,
German Speech Recognition: A Solution for the Analysis and Processing of Lecture Recordings
German Speech Recognition: A Solution for the Analysis and Processing of Lecture Recordings Haojin Yang, Christoph Oehlke, Christoph Meinel Hasso Plattner Institut (HPI), University of Potsdam P.O. Box
Predict Influencers in the Social Network
Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, [email protected] Department of Electrical Engineering, Stanford University Abstract Given two persons
Chapter 4: Artificial Neural Networks
Chapter 4: Artificial Neural Networks CS 536: Machine Learning Littman (Wu, TA) Administration icml-03: instructional Conference on Machine Learning http://www.cs.rutgers.edu/~mlittman/courses/ml03/icml03/
Develop Software that Speaks and Listens
Develop Software that Speaks and Listens Copyright 2011 Chant Inc. All rights reserved. Chant, SpeechKit, Getting the World Talking with Technology, talking man, and headset are trademarks or registered
Direct Loss Minimization for Structured Prediction
Direct Loss Minimization for Structured Prediction David McAllester TTI-Chicago [email protected] Tamir Hazan TTI-Chicago [email protected] Joseph Keshet TTI-Chicago [email protected] Abstract In discriminative
Dragon Solutions. Using A Digital Voice Recorder
Dragon Solutions Using A Digital Voice Recorder COMPLETE REPORTS ON THE GO USING A DIGITAL VOICE RECORDER Professionals across a wide range of industries spend their days in the field traveling from location
School Class Monitoring System Based on Audio Signal Processing
C. R. Rashmi 1,,C.P.Shantala 2 andt.r.yashavanth 3 1 Department of CSE, PG Student, CIT, Gubbi, Tumkur, Karnataka, India. 2 Department of CSE, Vice Principal & HOD, CIT, Gubbi, Tumkur, Karnataka, India.
Automatic Evaluation Software for Contact Centre Agents voice Handling Performance
International Journal of Scientific and Research Publications, Volume 5, Issue 1, January 2015 1 Automatic Evaluation Software for Contact Centre Agents voice Handling Performance K.K.A. Nipuni N. Perera,
BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION
BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION P. Vanroose Katholieke Universiteit Leuven, div. ESAT/PSI Kasteelpark Arenberg 10, B 3001 Heverlee, Belgium [email protected]
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING USING NEURAL NETWORKS JothiKumar.R 1, Sivabalan.R.V 2 1 Research scholar, Noorul Islam University, Nagercoil, India Assistant Professor, Adhiparasakthi College
How To Recognize Voice Over Ip On Pc Or Mac Or Ip On A Pc Or Ip (Ip) On A Microsoft Computer Or Ip Computer On A Mac Or Mac (Ip Or Ip) On An Ip Computer Or Mac Computer On An Mp3
Recognizing Voice Over IP: A Robust Front-End for Speech Recognition on the World Wide Web. By C.Moreno, A. Antolin and F.Diaz-de-Maria. Summary By Maheshwar Jayaraman 1 1. Introduction Voice Over IP is
Comparing Support Vector Machines, Recurrent Networks and Finite State Transducers for Classifying Spoken Utterances
Comparing Support Vector Machines, Recurrent Networks and Finite State Transducers for Classifying Spoken Utterances Sheila Garfield and Stefan Wermter University of Sunderland, School of Computing and
AUDIMUS.media: A Broadcast News Speech Recognition System for the European Portuguese Language
AUDIMUS.media: A Broadcast News Speech Recognition System for the European Portuguese Language Hugo Meinedo, Diamantino Caseiro, João Neto, and Isabel Trancoso L 2 F Spoken Language Systems Lab INESC-ID
Thirukkural - A Text-to-Speech Synthesis System
Thirukkural - A Text-to-Speech Synthesis System G. L. Jayavardhana Rama, A. G. Ramakrishnan, M Vijay Venkatesh, R. Murali Shankar Department of Electrical Engg, Indian Institute of Science, Bangalore 560012,
6.2.8 Neural networks for data mining
6.2.8 Neural networks for data mining Walter Kosters 1 In many application areas neural networks are known to be valuable tools. This also holds for data mining. In this chapter we discuss the use of neural
Web Document Clustering
Web Document Clustering Lab Project based on the MDL clustering suite http://www.cs.ccsu.edu/~markov/mdlclustering/ Zdravko Markov Computer Science Department Central Connecticut State University New Britain,
Advanced Signal Processing and Digital Noise Reduction
Advanced Signal Processing and Digital Noise Reduction Saeed V. Vaseghi Queen's University of Belfast UK WILEY HTEUBNER A Partnership between John Wiley & Sons and B. G. Teubner Publishers Chichester New
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
9RLFH$FWLYDWHG,QIRUPDWLRQ(QWU\7HFKQLFDO$VSHFWV
Université de Technologie de Compiègne UTC +(8',$6
A FUZZY BASED APPROACH TO TEXT MINING AND DOCUMENT CLUSTERING
A FUZZY BASED APPROACH TO TEXT MINING AND DOCUMENT CLUSTERING Sumit Goswami 1 and Mayank Singh Shishodia 2 1 Indian Institute of Technology-Kharagpur, Kharagpur, India [email protected] 2 School of Computer
Information Leakage in Encrypted Network Traffic
Information Leakage in Encrypted Network Traffic Attacks and Countermeasures Scott Coull RedJack Joint work with: Charles Wright (MIT LL) Lucas Ballard (Google) Fabian Monrose (UNC) Gerald Masson (JHU)
Offline Recognition of Unconstrained Handwritten Texts Using HMMs and Statistical Language Models. Alessandro Vinciarelli, Samy Bengio and Horst Bunke
1 Offline Recognition of Unconstrained Handwritten Texts Using HMMs and Statistical Language Models Alessandro Vinciarelli, Samy Bengio and Horst Bunke Abstract This paper presents a system for the offline
Dragon Solutions Using A Digital Voice Recorder
Dragon Solutions Using A Digital Voice Recorder COMPLETE REPORTS ON THE GO USING A DIGITAL VOICE RECORDER Professionals across a wide range of industries spend their days in the field traveling from location
Speech recognition for human computer interaction
Speech recognition for human computer interaction Ubiquitous computing seminar FS2014 Student report Niklas Hofmann ETH Zurich [email protected] ABSTRACT The widespread usage of small mobile devices
Sentiment analysis using emoticons
Sentiment analysis using emoticons Royden Kayhan Lewis Moharreri Steven Royden Ware Lewis Kayhan Steven Moharreri Ware Department of Computer Science, Ohio State University Problem definition Our aim was
A Content based Spam Filtering Using Optical Back Propagation Technique
A Content based Spam Filtering Using Optical Back Propagation Technique Sarab M. Hameed 1, Noor Alhuda J. Mohammed 2 Department of Computer Science, College of Science, University of Baghdad - Iraq ABSTRACT
Implementation of the Discrete Hidden Markov Model in Max / MSP Environment
Implementation of the Discrete Hidden Markov Model in Max / MSP Environment Paul Kolesnik and Marcelo M. Wanderley Sound Processing and Control Lab, Centre for Interdisciplinary Research in Music Media
Speech Signal Processing: An Overview
Speech Signal Processing: An Overview S. R. M. Prasanna Department of Electronics and Electrical Engineering Indian Institute of Technology Guwahati December, 2012 Prasanna (EMST Lab, EEE, IITG) Speech
Emotion in Speech: towards an integration of linguistic, paralinguistic and psychological analysis
Emotion in Speech: towards an integration of linguistic, paralinguistic and psychological analysis S-E.Fotinea 1, S.Bakamidis 1, T.Athanaselis 1, I.Dologlou 1, G.Carayannis 1, R.Cowie 2, E.Douglas-Cowie
How to write a technique essay? A student version. Presenter: Wei-Lun Chao Date: May 17, 2012
How to write a technique essay? A student version Presenter: Wei-Lun Chao Date: May 17, 2012 1 Why this topic? Don t expect others to revise / modify your paper! Everyone has his own writing / thinkingstyle.
Impact of Boolean factorization as preprocessing methods for classification of Boolean data
Impact of Boolean factorization as preprocessing methods for classification of Boolean data Radim Belohlavek, Jan Outrata, Martin Trnecka Data Analysis and Modeling Lab (DAMOL) Dept. Computer Science,
Gender Identification using MFCC for Telephone Applications A Comparative Study
Gender Identification using MFCC for Telephone Applications A Comparative Study Jamil Ahmad, Mustansar Fiaz, Soon-il Kwon, Maleerat Sodanil, Bay Vo, and * Sung Wook Baik Abstract Gender recognition is
Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics
Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics Anastasis Kounoudes 1, Anixi Antonakoudi 1, Vasilis Kekatos 2 1 The Philips College, Computing and Information Systems
VISUAL ALGEBRA FOR COLLEGE STUDENTS. Laurie J. Burton Western Oregon University
VISUAL ALGEBRA FOR COLLEGE STUDENTS Laurie J. Burton Western Oregon University VISUAL ALGEBRA FOR COLLEGE STUDENTS TABLE OF CONTENTS Welcome and Introduction 1 Chapter 1: INTEGERS AND INTEGER OPERATIONS
A secure face tracking system
International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 4, Number 10 (2014), pp. 959-964 International Research Publications House http://www. irphouse.com A secure face tracking
Voice Driven Animation System
Voice Driven Animation System Zhijin Wang Department of Computer Science University of British Columbia Abstract The goal of this term project is to develop a voice driven animation system that could take
Machine Translation. Agenda
Agenda Introduction to Machine Translation Data-driven statistical machine translation Translation models Parallel corpora Document-, sentence-, word-alignment Phrase-based translation MT decoding algorithm
Speech Recognition Software Review
Contents 1 Abstract... 2 2 About Recognition Software... 3 3 How to Choose Recognition Software... 4 3.1 Standard Features of Recognition Software... 4 3.2 Definitions... 4 3.3 Models... 5 3.3.1 VoxForge...
Things to remember when transcribing speech
Notes and discussion Things to remember when transcribing speech David Crystal University of Reading Until the day comes when this journal is available in an audio or video format, we shall have to rely
Improving Data Driven Part-of-Speech Tagging by Morphologic Knowledge Induction
Improving Data Driven Part-of-Speech Tagging by Morphologic Knowledge Induction Uwe D. Reichel Department of Phonetics and Speech Communication University of Munich [email protected] Abstract
Neural Networks algorithms and applications
Neural Networks algorithms and applications By Fiona Nielsen 4i 12/12-2001 Supervisor: Geert Rasmussen Niels Brock Business College 1 Introduction Neural Networks is a field of Artificial Intelligence
SIGN LANGUAGE CONVERTER
SIGN LANGUAGE CONVERTER Taner Arsan and Oğuz Ülgen Department of Computer Engineering, Kadir Has University, Istanbul, Turkey ABSTRACT The aim of this paper is to design a convenient system that is helpful
Objective Speech Quality Measures for Internet Telephony
Objective Speech Quality Measures for Internet Telephony Timothy A. Hall National Institute of Standards and Technology 100 Bureau Drive, STOP 8920 Gaithersburg, MD 20899-8920 ABSTRACT Measuring voice
SASSC: A Standard Arabic Single Speaker Corpus
SASSC: A Standard Arabic Single Speaker Corpus Ibrahim Almosallam, Atheer AlKhalifa, Mansour Alghamdi, Mohamed Alkanhal, Ashraf Alkhairy The Computer Research Institute King Abdulaziz City for Science
Emotion Detection from Speech
Emotion Detection from Speech 1. Introduction Although emotion detection from speech is a relatively new field of research, it has many potential applications. In human-computer or human-human interaction
Face Recognition For Remote Database Backup System
Face Recognition For Remote Database Backup System Aniza Mohamed Din, Faudziah Ahmad, Mohamad Farhan Mohamad Mohsin, Ku Ruhana Ku-Mahamud, Mustafa Mufawak Theab 2 Graduate Department of Computer Science,UUM
Performance Evaluation of Artificial Neural. Networks for Spatial Data Analysis
Contemporary Engineering Sciences, Vol. 4, 2011, no. 4, 149-163 Performance Evaluation of Artificial Neural Networks for Spatial Data Analysis Akram A. Moustafa Department of Computer Science Al al-bayt
CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREER-READY FOUNDATIONS IN ALGEBRA
We Can Early Learning Curriculum PreK Grades 8 12 INSIDE ALGEBRA, GRADES 8 12 CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREER-READY FOUNDATIONS IN ALGEBRA April 2016 www.voyagersopris.com Mathematical
Towards Full Comprehension of Swahili Natural Language Statements for Database Querying
6 Towards Full Comprehension of Swahili Natural Language Statements for Database Querying Lawrence Muchemi Natural language access to databases is a research area shrouded by many unresolved issues. This
Speech Recognition of a Voice-Access Automotive Telematics. System using VoiceXML
Speech Recognition of a Voice-Access Automotive Telematics System using VoiceXML Ing-Yi Chen Tsung-Chi Huang [email protected] [email protected] Department of Computer Science and Information
Supply Chain Forecasting Model Using Computational Intelligence Techniques
CMU.J.Nat.Sci Special Issue on Manufacturing Technology (2011) Vol.10(1) 19 Supply Chain Forecasting Model Using Computational Intelligence Techniques Wimalin S. Laosiritaworn Department of Industrial
2-1 Position, Displacement, and Distance
2-1 Position, Displacement, and Distance In describing an object s motion, we should first talk about position where is the object? A position is a vector because it has both a magnitude and a direction:
Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence
Artificial Neural Networks and Support Vector Machines CS 486/686: Introduction to Artificial Intelligence 1 Outline What is a Neural Network? - Perceptron learners - Multi-layer networks What is a Support
2014/02/13 Sphinx Lunch
2014/02/13 Sphinx Lunch Best Student Paper Award @ 2013 IEEE Workshop on Automatic Speech Recognition and Understanding Dec. 9-12, 2013 Unsupervised Induction and Filling of Semantic Slot for Spoken Dialogue
Text-To-Speech Technologies for Mobile Telephony Services
Text-To-Speech Technologies for Mobile Telephony Services Paulseph-John Farrugia Department of Computer Science and AI, University of Malta Abstract. Text-To-Speech (TTS) systems aim to transform arbitrary
Phone Recognition with the Mean-Covariance Restricted Boltzmann Machine
Phone Recognition with the Mean-Covariance Restricted Boltzmann Machine George E. Dahl, Marc Aurelio Ranzato, Abdel-rahman Mohamed, and Geoffrey Hinton Department of Computer Science University of Toronto
Slovak Automatic Transcription and Dictation System for the Judicial Domain
Slovak Automatic Transcription and Dictation System for the Judicial Domain Milan Rusko 1, Jozef Juhár 2, Marian Trnka 1, Ján Staš 2, Sakhia Darjaa 1, Daniel Hládek 2, Miloš Cerňak 1, Marek Papco 2, Róbert
Specialty Answering Service. All rights reserved.
0 Contents 1 Introduction... 2 1.1 Types of Dialog Systems... 2 2 Dialog Systems in Contact Centers... 4 2.1 Automated Call Centers... 4 3 History... 3 4 Designing Interactive Dialogs with Structured Data...
An Introduction to Data Mining
An Introduction to Intel Beijing [email protected] January 17, 2014 Outline 1 DW Overview What is Notable Application of Conference, Software and Applications Major Process in 2 Major Tasks in Detail
ACCESS 2007. Importing and Exporting Data Files. Information Technology. MS Access 2007 Users Guide. IT Training & Development (818) 677-1700
Information Technology MS Access 2007 Users Guide ACCESS 2007 Importing and Exporting Data Files IT Training & Development (818) 677-1700 [email protected] TABLE OF CONTENTS Introduction... 1 Import Excel
Neural Network based Vehicle Classification for Intelligent Traffic Control
Neural Network based Vehicle Classification for Intelligent Traffic Control Saeid Fazli 1, Shahram Mohammadi 2, Morteza Rahmani 3 1,2,3 Electrical Engineering Department, Zanjan University, Zanjan, IRAN
Achieving Success Through Assistive Technology
Achieving Success Through Assistive Technology Assistive Technology refers to any technology that allows one to increase, maintain, or improve the functional capabilities of an individual with special
Recent advances in Digital Music Processing and Indexing
Recent advances in Digital Music Processing and Indexing Acoustics 08 warm-up TELECOM ParisTech Gaël RICHARD Telecom ParisTech (ENST) www.enst.fr/~grichard/ Content Introduction and Applications Components
THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS
THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS O.U. Sezerman 1, R. Islamaj 2, E. Alpaydin 2 1 Laborotory of Computational Biology, Sabancı University, Istanbul, Turkey. 2 Computer Engineering
