Conclusion and Discussion will be presented in Section 5.
|
|
|
- Abigail Knight
- 9 years ago
- Views:
Transcription
1 Developing Speech Recognition System for Quranic Verse Recitation Learning Software B.Putra, B.T. Atmaja, D.Prananto 3 Department of Engineering Physics Sepuluh opember Institutes of Technology 60 Surabaya, Indonesia { budiman.fgi, btatmaja, 3 dprananto}@gmail.com Abstract Quran as holy book for Muslim consists of many rules which are needed to be considered in reading Quran verse properly. If the recitation does not meet all of those rules, the meaning of Quran verse recited will be different with its origins. Intensive learning is needed to be able to do correct recitation. However, the limitation of teachers and time to study Quran verse recitation together in a class could be an obstacle in Quran recitation learning. In order to minimize the obstacle and to ease the learning process we implement speech recognition techniques based on Mel Frequency Cepstral Coefficient (MFCC) features and Gaussian Mixture Model (GMM) modeling, we have successfully designed and developed Quran verse recitation learning software in prototype stage. This software is interactive multimedia software which has many features for learning flexibility and effectiveness. This paper explains the developing of speech recognition system for Quran learning software which is built with the ability to perform evaluation and correction in Qur an recitation. In this paper, the authors present clearly the built and tested prototype of the system based on experiment data. Keywords-Qur an verse recitation, speech recognition, Mel Frequency Cepstral Coefficient (MFCC), Gaussian Mixture Mode (GMM)l. I. ITRODUCTIO Al Qur an is the holy book of Moslem consisted of life guidance which has to be understood clearly by them. Because of that, it is important for the Moslem to be able recite Qur an verse correctly. Unfortunately, that is not easy because the Qur an recitation has many rules which have to be strictly considered on the whole words. Moreover, if the recitation does not meet the rules, the recitation will give different meaning from the origins. So, the Moslems have to learn intensively in order to be capable in reciting Qur an verse properly. recitation is aimed to ease people in studying how to recite Quranic verses by them with correct and proper manner without abandoning the conventional way, the face to face method. There are some methods which can be used to design and develop learning software for Quranic verse recitation. In the previous research which has done by Razak [4], the implementation of speech recognition for Quranic utterance style recognition using Hidden Markov Model (HMM) was developed. In this research, we use speech recognition techniques as part of template referencing method to develop the software. The techniques include Mel Frequency Cepstral Coefficient (MFCC) and Gaussian Mixture model (GMM) which use the threshold of log-likelihood value as the evaluation of utterances. This software will carry out correction and rectification of reading on each learning session if there is any error in reading. It is expected that people could learn to read the Quran easily without feeling doubt about the accuracy of pronunciation and recitation of Quranic verse recitation law (tajweed), because they would feel to have a virtual mentor which always makes corrections in learning process. This paper is a revised version of conference paper presented at International Conference of Informatics for Development held Yogyakarta, Indonesia. Some of parts are modified and expanded from the origins. The speech recognition theory explained is revised to be more comprehensive with the research and more detail from the origins. Then, the research methodology part is modified with major changes to be clearer than before. In that part, we add the user interfaces of software as subsection to give the illustration for reader about our prototype software looks like and features built in this research. In addition, some minor revision also has done in other parts such as introduction part and the Quranic laws and pronunciations parts. In contrast, the conventional learning method for Qur an The rest of this paper is organized as follows. Section verse recitation learning is still uses face to face method which presents the theory of Quranic verse recitation law and the student has to be taught directly by the teacher. Limited pronunciation. The next section will explain speech recognition time to study together in a class and also the limited number of techniques and its sub-theory which is necessary in building teachers result in lack learning process. Then the other this system. The implementation and configuration on the learning lack process is the absence of lecturer because in building stage of software are described in the next section in conventional method, the lecture holds the important role in Research Methodology. Then, the experiment conducted and there. The development of learning software for Qur an verse its result is explained in Experiment Result section. Finally, the Conclusion and Discussion will be presented in Section 5. This paper is a revised and expanded version of a paper entitled Prototyping of Quranic Verse Recitation Learning Software Using Speech Recognition ISS Techniques : 5 Based on Cepstral Feature presented at International Conference 4 on Informatics for Development 0, Yogyakarta, Indonesia []
2 II. A. Makhraj QURAIC VERSE RECITATIO LAW AD PROOUCIATIO Makhraj means place of discharge. Makhraj in hija iyah letters means the place where hija iyah letters come out from mouth. Considering that Arabic letters differ from latin letters, hence Arabic letters pronounced in different manner to malay words pronunciation. Pronunciation of Arabic words/letter is determined by makhraj of the letter. B. Mad Mad means elongate tone. In tajweed course there are two kind of mad, i.e. mad ashli/tabi i and mad far i. Ashli means principal and far i means subsidiary. Mad ashli is read in two harakat. Mad ashli occur when there is alif after letters with fathah, ya consonant after letters with kasrah, and waw consonant after letters with dhomah. Furthermore, long-sign readings can be utilized if alif, waw consonant or ya consonant are not used. Mad far i has many different types with different vowel length. The application of the Mad is also different for every qiraat. C. Law of un Consonant The law of nun consonant and tanween can be classified in some types. Izhhaar means clearly read. Izhaar is readings where nun consonant or tanween meet alif ( ), hamzah ( ), ain ( ), ghain ( ), ha ( ), kha ( ), and ha ( ) and read with clear sound. Idgham means to get into/to change the tone of nun when nun consonant or tanween meet idgham letters. Each idgham readings read in two harakat. Idgham Letters are ya ( ) ra ( ), mim ( ), lam ( ), waw ( ), and nun ( ). Idgham bilaghunnah is an inverse of idgham bighunnah, where the tone is not inserted into the nose. The letters of idgham bilaghunnah are lam and ra. Iqlab occur when nun consonant or tanween meet ba ( ). The tone of ba in Iqlab readings changed to mim accompanied with drone. Iqlab readings are read in two harakat. Ikhfa means to hide/vague. Ikhfa readings is read by vague voice of nun when nun consonant or tanween meet ikhfa letters. All the readings with ikhfa read in two harakats. Ikhfa letters are the letters except in izhhaar, idgham and iqlab. D. Law Mimi Consonant Idgam mutamatsilayn occurs when mim consonant meet mim, where the tone of mim in mim consonant is inserted to the next letter tone with buzz tone. This readings is read with two harakats. If mim consonant meet ba, then the tone of mim in mim consonant read with vague with a bit of buzz. It is read with two harakats. If mim consonant meet letters other than mim and ba, then the tone of mim in mim consonant read obviously. Izhaar syafawi is read with one harakat. Ghunnah means to buzz. Ghunnah occur in two cases, that is when mim and nun use tasydid sign. Ghunnah is read with two harakats. III. SPEECH RECOGITIO TECHIQUES Signal processing was utilized to obtain the characteristics of pronunciation which latter be used as identifier of faults and correction. Extraction of MFCC coefficient, signal energy, delta MFCC and delta-delta MFCC is conducted to subtract the feature of voices. A. Voice Signal and its Occurrence Process Voice is signals which greatly influenced by frequency and a form of discrete signal which is influenced by time. The main component in voice production system is vocal tract. Vocal tract is a resonance tube-shaped object in voice production system which has three main parts called pharynx, nasal cavity and oral cavity. The vocal tract varies in shapes according to soft plate (velum), tongues, lips and jaw which overall called as articulators. See Figure below, the vocal organ consists of () asal cavity, () Hard palate, (3) Alveolar ridge, (4) Soft palate (Velum), (5) Tip of the tongue (Apex), (6) Dorsum, (7) Uvula, (8) Radix, (9) Pharynx, (0) Epiglottis, () False vocal cords, () Vocal cords, (3) Larynx, (4) Esophagus, and (5) Trachea. Figure. Human vocal organs [4] The process of human voice production can be explained as follows; the air flows from lungs to the trachea, a tube composed of cartilage rings, and goes through larynx to the vocal tract. Larynx reacts as a gate between lungs and mouth. Larynx composed of epiglottis, vocal cords and false vocal cords. These three components are closed when human shallow food, so that the food does not enter the lungs, and open again when the human inhale. Phoneme in English can ISS :
3 be classified in terminology of manner of articulation and place of articulation. Manner of articulation is concentrated on air flow, it means it concerns about track and the level of vocal that goes trough. Manner of articulation and voicing is divided into three big classes of phoneme. Phoneme that produce employ voicing and solely stimulate the vocal tract on glottis called sonorant (vowels, diphthongs, glides, liquids and nasals). They have continuous, intents and periodic phonemes characteristics. Voiced sounds is produced by air pressure which flows trough vocal cords while vocal cords squeezed to open and close quickly to produce a series of puffs periodic which have fundamental frequency (first harmonics) same as vocal-cord s vibration frequency. The frequency of vocal cords depends on the level of solidity, tension, length of vocal cords and air flow effect which produced in the glottis, a chamber between vocal cords. The component of this frequency is composed of a number of harmonics from fundamental frequency. A sound that produced without vibration in vocal cords is called unvoiced []. The following picture illustrates a segment on vowel /ix/. A quasi-periodicity signal on voiced speech can be seen here. Affricate is a combination of stop and fricative sounds. Stops, fricative and affricates collectively called as obstruent phoneme which is weak enough and periodic, and basically is a form that generated by blocking stimulus on main vocal tract. Vowel is basically described in terminology of tongue position and lips B. Speech Signal Processing Speech signal processing is intended to obtain cepstral feature of human voice. The process of voice signal processing consists of Sampling, Frame Blocking, Windowing, Discrete Fourier Transform (DFT), Filter Bank, Discrete Cosine Transform (DCT) and calculating of dynamical coefficient and spectrum energy. ) Sampling Human voices will generate continuous analog signals. Therefore, the analog signal is chopped in certain interval of time. Discrete series sample x [n] is obtained from continuous signal x(t), xn [ ] = xnt ( ) () Where T is sampling period and i/t = Fs is sampling frequency in unit of sample/second. The value of n is the number of samples. According to the yquist sampling theory, minimal sampling frequency required is twice of original maximal signal. Figure. Illustration of vowel segment /ix/ [] FSampling FSignal () Fricative sound is generated from the confinement of vocal tract and air flow pressure with high enough velocity to make turbulence. The turbulence is for instance /hh/ or /s/. Figure 3. Fricative /s/ [] Plosive or stop sounds is generated from vocal tract blocking process by closing the lips and nasal cavity, enabling lateral air pressure, and followed by a beat. This mechanism will generate /p/ and /g/ voices. ) Frame Blocking Frame Blocking is segregation of voices into several frames where one frame consisted of several samples. This process is needed to transform a non-stationer signal to a quasi-stationer signal so it can be transformed from time domain to frequency domain with Fourier transform. Human voice signal indicate quasi-stationer characteristic in range of time from 0 to 40 milliseconds. Hence, in that range of time the Fourier transform can be performed. 3) Windowing Voice signal which is chopped into frames will lead to discontinuity in initial and final signal. The discontinuity leads to data error in the process of Fourier transform. Windowing is needed to reduce the effect of discontinuity in chopped signal. If window is defined as w(n), where 0 n - and is number of samples in each frame, then the result of windowing process is; wn ( ) = xnw ( ) ( n), 0 n (3) Figure 4. Plosive or stop sounds /g/ [] ISS :
4 Windowing which is used in this research is Hamming windowing. To obtain the MFCC, the result of energy logarithmic is processed with Discrete Cosine Transform (DCT). ( cos(π n /( )) 0 n (4) W hann = 0 others [ ] c n M π nm ( 0.5) s[ m]cos m 0 M = = (9) 4) Discrette Fourier Transform (DFT) Fourier transform is performed to transform from time domain to frequency domain. DFT is a specific form of integral Fourier equation; Y( ω) = w( t) e jω dt (5) The DFT can be obtained by changing the variables time (t) and frequency (w) into discrete form: jkω ( ) ( ) 0 nt Y kω 0 = w nt e (6) n = 0 5) Mel Frequency Cepstral Coefficient (MFCC) The most important information of human voice signal is located at high frequency. This important information determines the characteristic of human voice and Mel Scale is utilized to accommodate this characteristics. The relation between Mel and actual frequency is according to various researches about perception of voice reception by human ear. Mel = 000 log ( + ω) (7) In the implementation, this scaling is interpreted with Mel Filter Bank where each value of frequency magnitude is filtered by triangle filter series with Mel frequency as middle frequency. The triangular filter represents the process of Mel scaling in the signal. 6) Calculation of Dynamical Cepstral Coefficient and Spectrum Energy. In order to add the pattern of recitation speech, the system is equipped with dynamical cepstral coefficient and spectrum energy. The dynamical cepstral coefficient is inherited from MFCC each frame separated to two forms coefficient, delta MFCC and delta-delta MFCC Delta MFCC naturally is the first regression of MFCC in each frame. The equation below is the computation of delta MFCC. d n( cn+ n= i = n= n c n ) (0) Where d i is the delta coefficient at frame i computed in terms of the corresponding basic coefficients c n+ to c n. The same equation is used to compute the acceleration coefficients by replacing the basic coefficients with the delta coefficients. Then, the delta-delta MFCC or acceleration of MFCC can be computed as follows, a n( dn+ n= i = n= n d n ) () where a i is the delta coefficient at frame i computed in terms of the corresponding basic coefficients d n+ to d n Figure 5. Mel filter bank construction After the magnitude of signal spectrum X[k] filtered by Mel Filter Bank, computation of logarithmic value of energy is conducted to each of output band from each filter. Logarithmic signal energy process is utilized to adapt the system just like human ear. s m = [ ] ln k= 0 (8) m [ Y[ k ] H [ k] 0 m M C. Gaussian Mixture Model (GMM) Gaussian probability density function (pdf) is bellshaped one dimensional function which is defined by two parameters, that is mean µ and variant σ or covariant Σ. In D dimension it can be formulated as; T ( x;µ, ) = exp ( x µ ) ( x µ ) D/ / ( π ) () ISS :
5 where µ is vector mean and Σ is covariant matrix. Gaussian mixture model (GMM) is a mixing of several Gaussian distribution or representation of the existence of subclasses in a class. The probability density function of GMM is described as sum of multiplication of weight with Gaussian probability. C p ( x; θ) = α c ( x;µ c, c ) c = α c is weight of mixed component c, where 0 < each component distribution, C and α c. c = (3) αc < for Whereas, the parameter θ = {α, µ,,..., α C, µc, C } (4) It is the definition of Gaussian mixture probability density function parameter. The figure blow is the illustration of GMM with three mixtures kind of recitation law such as Idhar, Ikhfa or Iqlab only. The highest level, advance, include courses with combination of more than one law of recitation and still with pronunciation correction. Moreover in this last level, both pronunciation and verse recitation law are evaluated. So, the higher level we take the higher difficulty of the course. If user makes an error during the learning, the error message will appear on the software which will guide the user to correct the reading and pronunciations. In order to support the learning process, each module in the software equipped with the sample recitation. This is needed to make the user understand how to recite correctly. Also by this recitation sample, the user can copy the recitation to ease his learning process. As the software supporting features, software is also equipped with a tutorial how to read Al-Qur an recitation correctly as guidance and pre-learning. The tutorial is recommended to read every time we make error so we know why it is wrong. B. Design of Correction System The figure below shows the general step of designing and application of software, Figure 6. Example of GMM dimension- with 3 mixtures IV. RESEARCH METHODOLOGY This part will explain about the prototype system and the testing procedure to obtain the system performance. Also, in the last of this part the result of recognition performance test is reported. A. Description of software and features This software is a prototype stage software used to learn how to read Al Qur an correctly. Commonly, Al-Quran recitation learning book require more competent supervisor. However, this software is an interactive multimedia software accompanied with pronunciation and verse recitation law correction in each courses. This software consists of three modules or levels depend of the difficulty and learning material, i.e. basic, intermediate and advance. In the basic level, the course includes only correction in makhraj. The module included in this level consists of letters of Hijaiyah. Then, in intermediate level, the courses consist of law of recitation and also correction in makhraj/pronunciation. In this level there include only one ISS : (a) (b) Figure 7. (a) Designing and template making step of software using expert/ustadz recitation (b) Process of software application used by user The basic idea of the correction and evaluation system is template matching using speech recognition, which the cepstral feature of the voice of each reading is taken for main pattern of recognition. The voice used as main pattern is recorded from the expert of Qur an recitation. Also this voice can be played in software as module recitation sample voice. The Gaussian Mixture Model (GMM) is used to model the speech features of recitation. In this research, the number of mixture in each GMM consisted of 6 mixtures. This number is selected based on consideration of minimum software computation time. 8
6 Formulation and design is done prior to software construction and design. General formulation of the parameters of GMM modeling which is used to recognize the reading is obtained from this stage. The stage of design and formulation is composed of digital signal processing (DSP), GMM modeling and logging the GMM parameters into software database. Digital signal processing is consisted of several sub process as follows, see Figure 8. The incoming voice signal is processed in order to extract the cepstral feature parameters. The incoming voice signal is sampled with sampling frequency of 8000 Hz, accordance with yquist rule, and then divided into time slots (framing) with frame time length 40 ms and overlapping time 0 ms or about 50%. The number of frame is separated depend on the recitation word number or harakat. Then, each frame is passed through Hamming window to reduce signals discontinuity after framing. The signal is then transformed to frequency domain by using DFT with = 04. The signal is passed through the Filter bank of 4 triangles mel filter. The DCT with MFCC coefficient of 4 is done after filtering process. Other feature calculation like signal energy, delta- MFCC and delta-delta MFCC were done after DCT process. Figure 9. Process of Software Construction After database is filled by GMM parameters, the software can be used in application or software using stage. At this stage, correction process is done with similar process at the construction and formulation stage. However, as figure 7b on the top, at this stage likelihood estimation with GMM parameter for each recitation law/pronunciation is done after signal processing process. Furthermore, if the likelihood value is less than the predetermined threshold value then certainly an error occurred in reciting the readings, see figure 0. The error message a is appeared in order to warn the user to reread the readings with proper and correct law of recitation. Figure 0. The Correction Process of Software C. User Interfaces (UI) As user software, the interface is important part attracting the user to use the software. Below is the user interface of software we built in three level modules, Figure 8. Digital Signal Processing for Extracting Cepstral Feature The four signal feature vectors in each group word are used as input of the observation data in the GMM. The number of state and model which is used to design the software is not specified to adapt with verses section or the session of each reading in the course. The process of GMM modeling can be seen in figure below. The first step is initialization of clustering as many as number of GMM mixtures. After initialization by K-Means clustering, the next process is GMM training to obtain the maximum likelihood using Expectation-Maximation (EM) algorithm. In this training use maximum iteration until 000 and convergence value is After several iterations, if convergence or maximum iteration is achieved the iteration can be stopped and all the parameter is stored into the software database. Each recitation law correlated with word numbers (harakat) in the course has one GMM model for correction as the reference template of recitation model in the application stage. (a) ISS :
7 (b) (c) (d) Figure. User infeace software (a) in standby mode (b) level Basic Module (c) level Intermediate Module (d) level Advance Module We can see figure on the top the user interface of this prototype software is designed simply. In further research, we develop this interface in object to ease the user in learning process. The accuracy of correction for hija iyah letter pronunciation is obtained from the average result for all hija iyah letters. For example, how much alif recited true by the system compared with real true value of pronunciation of alif and then other hija iyah letters. The average result from all hija iyah letters is calculated as the correction accuracy which the value is 90%. This value is higher than we expected. However, this value is premature data as the experiment data is not enough for evaluate the speech recognition based system. ) Accuracy of Quranic Recitation Law Accuracy of Quranic recitation law can be obtained by testing the system of Quranic recitation. For a word containing some letters, the law might be idghom, ihkfa and idhar. How many law from some Arabic words is detected true compared with real true value is obtained as the accuracy of Quranic recitation law. For recitation law correction test, the result suggests that the correction accuracy is fairly poor (70%), hence, the system needs a reconfiguration in order to improve the correction of recitation law. 3) Combination of several Makhraj and recitation law The last accuracy test for the system is the combination of makhraj and Quranic recitation law. It can be obtained by testing both makhraj and Quranic recitation law from some Arabic words and compare the result with real true value. For combination of several makhraj and recitation law test, the result suggests that the correction accuracy is also poor (60%), hence, the system need a reconfiguration back in order to improve combination of makhraj and recitation law. VI. COCLUSIO The prototype has been designed as an interactive multimedia Quran recitation learning software using cepstral feature and GMM modeling as the basis of speech recognition technology. The results of the research suggest that the correction accuracy of the software is 70% for pronunciation, 90% for recitation law and 60% for combination of pronunciation and recitation law. The performance of correction accuracy can be improved by changing the configuration and template of speech recognition used.. We are working on adding the speech enhancement and data training in object to improve the performance of learning software. Also, recently we are developing the features of software in order to improve the feasibility of this software to be used. REFERECES V. RESULT AD DISCUSSIO In order to test the reliability and accuracy of correction, experiment for each recitation law in the software is done with ten speakers to read the readings in wrong and right manner. This test used clean speech signal as the template. [] B.Putra, B.T. Atmaja, D.Parananto. 0. Prototyping of Quranic Verse Recitation Learning Software Using Speech Recognition Techniques Based on Cepstral Feature. Proceeding of International Conference of Informatics Development 6 overmber 0, Yogyakarta, Indonesia. [] Bardici, ick. Speech Recognition using Gaussian Mixture Model. PhD Thesis. Blekinge Institute of Technology. 006 [3] Moreno, Pedro J. Speech Recognition in oisy Environments. PhD Thesis. Carnegie Mellon University. 996 ) Analysis of pronunciation correction (Makhraj) ISS :
8 [4] Razak, Zaidi. Quranic Verse Recitation Feature Extraction Using Mel- Frequency Cepstral Coefficient (MFCC). Journal University of Malaya. 007 [5] Berouti, M., Schwartz, R. And Makhoul, J.. Enhancement Of Speech Corrupted By Acoustic oise. Proc. Of IEEE ICASSP, Pp. 08-, Washington DC.979 [6] Bhatnagar, B.E., Mukul. A Modified Spectral Subtraction Method Combined With Perceptual Weighting For Speech Enhancement. MSc Thesis. The University Of Texas At Dallas. 00 [7] Guard, Cedric. Speech Recognition Based On Template Matching and Phone Posteriori Probability. MSc Thesis. IDIAP Research Institute [8] Patel, Ibrahim. Speech Recognition Using HMM With MFCC-An Analysis Using Frequency Specral Decomposion Technique. Signal & Image Processing : An International Journal(SIPIJ) Vol., o., December 00. [9] Lima, Carlos. Spectral ormalisation MFCC Derived Features for Robust Speech Recognition. SPECOM 004: 9 th Conference Speech and Computer St. Petersburg, Russia September 0-, 004. [0] Bala, Anjali. Voice Command Recognition System Based On MFCC And DTW. International Journal of Engineering Science and Technology Vol. (), 00, [] Shaneh, Mahdi. Voice Command Recognition System Based on MFCC and VQ Algorithms. World Academy of Science, Engineering and Technology 57, 009. [] Rabiner, Lawrence R. A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition. Proceedings of the IEEE Vol 77 o, February 989. [3] B.H. Juang, Lawrence R Rabiner. Hidden Markov Model for Speech Recognition. Technometrics, Vol. 33, o. 3. (Aug., 99), pp [4] Review of Speech Synthesis Technology. files/theses/lemmetty_mst/index.html. [Online]. Available: publications/files/theses/ lemmetty_mst/chap3.html AUTHORS PROFILE Budiman Putra received his Bachelor of Engineering in Engineering Physics majoring Instrumentation and Control System Engineering from Department of Engineering Physics, Sepuluh opember Institute of Technology Surabaya, Indonesia. Currently, he is a Defense and Control System Engineer in PT Len Industri (Persero), Bandung, Indonesia. His research interests include Digital Signal Processing, Image Processing, Artificial Intelligence and Electronic Control System. Bagus Tris Atmaja received his Bachelor and Master of Engineering degree from Department of Engineering Physics, Sepuluh opember Institutes of Technology in 009 and 0 respectively. ow, he is a research sudent in Usagawa-Chisaki Lab, Kumamoto Univ., Japan. His research interests include Artificial Intelligence, Digital Signal Processing, Acoustics and Vibration. Dwi Prananto received his Bachelor degree in Engineering Physic from Sepuluh opember Institute of Technology (ITS), Surabaya, Indonesia. Currently, he is Master Student at Department of Physics, Tohoku University, Japan. His research interests include Smart Material, Thermoelectric Materials and Instrumentation Scientific. ISS :
L9: Cepstral analysis
L9: Cepstral analysis The cepstrum Homomorphic filtering The cepstrum and voicing/pitch detection Linear prediction cepstral coefficients Mel frequency cepstral coefficients This lecture is based on [Taylor,
SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA
SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA Nitesh Kumar Chaudhary 1 and Shraddha Srivastav 2 1 Department of Electronics & Communication Engineering, LNMIIT, Jaipur, India 2 Bharti School Of Telecommunication,
Developing an Isolated Word Recognition System in MATLAB
MATLAB Digest Developing an Isolated Word Recognition System in MATLAB By Daryl Ning Speech-recognition technology is embedded in voice-activated routing systems at customer call centres, voice dialling
School Class Monitoring System Based on Audio Signal Processing
C. R. Rashmi 1,,C.P.Shantala 2 andt.r.yashavanth 3 1 Department of CSE, PG Student, CIT, Gubbi, Tumkur, Karnataka, India. 2 Department of CSE, Vice Principal & HOD, CIT, Gubbi, Tumkur, Karnataka, India.
Automatic Evaluation Software for Contact Centre Agents voice Handling Performance
International Journal of Scientific and Research Publications, Volume 5, Issue 1, January 2015 1 Automatic Evaluation Software for Contact Centre Agents voice Handling Performance K.K.A. Nipuni N. Perera,
Lecture 12: An Overview of Speech Recognition
Lecture : An Overview of peech Recognition. Introduction We can classify speech recognition tasks and systems along a set of dimensions that produce various tradeoffs in applicability and robustness. Isolated
Hardware Implementation of Probabilistic State Machine for Word Recognition
IJECT Vo l. 4, Is s u e Sp l - 5, Ju l y - Se p t 2013 ISSN : 2230-7109 (Online) ISSN : 2230-9543 (Print) Hardware Implementation of Probabilistic State Machine for Word Recognition 1 Soorya Asokan, 2
Ericsson T18s Voice Dialing Simulator
Ericsson T18s Voice Dialing Simulator Mauricio Aracena Kovacevic, Anna Dehlbom, Jakob Ekeberg, Guillaume Gariazzo, Eric Lästh and Vanessa Troncoso Dept. of Signals Sensors and Systems Royal Institute of
Speech recognition for human computer interaction
Speech recognition for human computer interaction Ubiquitous computing seminar FS2014 Student report Niklas Hofmann ETH Zurich [email protected] ABSTRACT The widespread usage of small mobile devices
MUSICAL INSTRUMENT FAMILY CLASSIFICATION
MUSICAL INSTRUMENT FAMILY CLASSIFICATION Ricardo A. Garcia Media Lab, Massachusetts Institute of Technology 0 Ames Street Room E5-40, Cambridge, MA 039 USA PH: 67-53-0 FAX: 67-58-664 e-mail: rago @ media.
COMPARATIVE STUDY OF RECOGNITION TOOLS AS BACK-ENDS FOR BANGLA PHONEME RECOGNITION
ITERATIOAL JOURAL OF RESEARCH I COMPUTER APPLICATIOS AD ROBOTICS ISS 2320-7345 COMPARATIVE STUDY OF RECOGITIO TOOLS AS BACK-EDS FOR BAGLA PHOEME RECOGITIO Kazi Kamal Hossain 1, Md. Jahangir Hossain 2,
Emotion Detection from Speech
Emotion Detection from Speech 1. Introduction Although emotion detection from speech is a relatively new field of research, it has many potential applications. In human-computer or human-human interaction
Articulatory Phonetics. and the International Phonetic Alphabet. Readings and Other Materials. Introduction. The Articulatory System
Supplementary Readings Supplementary Readings Handouts Online Tutorials The following readings have been posted to the Moodle course site: Contemporary Linguistics: Chapter 2 (pp. 15-33) Handouts for This
Myanmar Continuous Speech Recognition System Based on DTW and HMM
Myanmar Continuous Speech Recognition System Based on DTW and HMM Ingyin Khaing Department of Information and Technology University of Technology (Yatanarpon Cyber City),near Pyin Oo Lwin, Myanmar Abstract-
Establishing the Uniqueness of the Human Voice for Security Applications
Proceedings of Student/Faculty Research Day, CSIS, Pace University, May 7th, 2004 Establishing the Uniqueness of the Human Voice for Security Applications Naresh P. Trilok, Sung-Hyuk Cha, and Charles C.
4 Phonetics. Speech Organs
4 Phonetics Speech is a very hierarchical and complex physical phenomenon, including issues related to cognition, language, physiology, hearing and acoustics. A research including features of these fields
A Sound Analysis and Synthesis System for Generating an Instrumental Piri Song
, pp.347-354 http://dx.doi.org/10.14257/ijmue.2014.9.8.32 A Sound Analysis and Synthesis System for Generating an Instrumental Piri Song Myeongsu Kang and Jong-Myon Kim School of Electrical Engineering,
Speech Signal Processing: An Overview
Speech Signal Processing: An Overview S. R. M. Prasanna Department of Electronics and Electrical Engineering Indian Institute of Technology Guwahati December, 2012 Prasanna (EMST Lab, EEE, IITG) Speech
Artificial Neural Network for Speech Recognition
Artificial Neural Network for Speech Recognition Austin Marshall March 3, 2005 2nd Annual Student Research Showcase Overview Presenting an Artificial Neural Network to recognize and classify speech Spoken
Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics
Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics Anastasis Kounoudes 1, Anixi Antonakoudi 1, Vasilis Kekatos 2 1 The Philips College, Computing and Information Systems
Lecture 1-10: Spectrograms
Lecture 1-10: Spectrograms Overview 1. Spectra of dynamic signals: like many real world signals, speech changes in quality with time. But so far the only spectral analysis we have performed has assumed
FFT Algorithms. Chapter 6. Contents 6.1
Chapter 6 FFT Algorithms Contents Efficient computation of the DFT............................................ 6.2 Applications of FFT................................................... 6.6 Computing DFT
The Calculation of G rms
The Calculation of G rms QualMark Corp. Neill Doertenbach The metric of G rms is typically used to specify and compare the energy in repetitive shock vibration systems. However, the method of arriving
Solutions to Exam in Speech Signal Processing EN2300
Solutions to Exam in Speech Signal Processing EN23 Date: Thursday, Dec 2, 8: 3: Place: Allowed: Grades: Language: Solutions: Q34, Q36 Beta Math Handbook (or corresponding), calculator with empty memory.
Separation and Classification of Harmonic Sounds for Singing Voice Detection
Separation and Classification of Harmonic Sounds for Singing Voice Detection Martín Rocamora and Alvaro Pardo Institute of Electrical Engineering - School of Engineering Universidad de la República, Uruguay
Statistical Machine Learning from Data
Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Gaussian Mixture Models Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique
BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION
BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION P. Vanroose Katholieke Universiteit Leuven, div. ESAT/PSI Kasteelpark Arenberg 10, B 3001 Heverlee, Belgium [email protected]
APPLYING MFCC-BASED AUTOMATIC SPEAKER RECOGNITION TO GSM AND FORENSIC DATA
APPLYING MFCC-BASED AUTOMATIC SPEAKER RECOGNITION TO GSM AND FORENSIC DATA Tuija Niemi-Laitinen*, Juhani Saastamoinen**, Tomi Kinnunen**, Pasi Fränti** *Crime Laboratory, NBI, Finland **Dept. of Computer
Speech: A Challenge to Digital Signal Processing Technology for Human-to-Computer Interaction
: A Challenge to Digital Signal Processing Technology for Human-to-Computer Interaction Urmila Shrawankar Dept. of Information Technology Govt. Polytechnic, Nagpur Institute Sadar, Nagpur 440001 (INDIA)
Advanced Signal Processing and Digital Noise Reduction
Advanced Signal Processing and Digital Noise Reduction Saeed V. Vaseghi Queen's University of Belfast UK WILEY HTEUBNER A Partnership between John Wiley & Sons and B. G. Teubner Publishers Chichester New
Final Year Project Progress Report. Frequency-Domain Adaptive Filtering. Myles Friel. Supervisor: Dr.Edward Jones
Final Year Project Progress Report Frequency-Domain Adaptive Filtering Myles Friel 01510401 Supervisor: Dr.Edward Jones Abstract The Final Year Project is an important part of the final year of the Electronic
LOW COST HARDWARE IMPLEMENTATION FOR DIGITAL HEARING AID USING
LOW COST HARDWARE IMPLEMENTATION FOR DIGITAL HEARING AID USING RasPi Kaveri Ratanpara 1, Priyan Shah 2 1 Student, M.E Biomedical Engineering, Government Engineering college, Sector-28, Gandhinagar (Gujarat)-382028,
Analysis/resynthesis with the short time Fourier transform
Analysis/resynthesis with the short time Fourier transform summer 2006 lecture on analysis, modeling and transformation of audio signals Axel Röbel Institute of communication science TU-Berlin IRCAM Analysis/Synthesis
The Algorithms of Speech Recognition, Programming and Simulating in MATLAB
FACULTY OF ENGINEERING AND SUSTAINABLE DEVELOPMENT. The Algorithms of Speech Recognition, Programming and Simulating in MATLAB Tingxiao Yang January 2012 Bachelor s Thesis in Electronics Bachelor s Program
T-61.184. Automatic Speech Recognition: From Theory to Practice
Automatic Speech Recognition: From Theory to Practice http://www.cis.hut.fi/opinnot// September 27, 2004 Prof. Bryan Pellom Department of Computer Science Center for Spoken Language Research University
The continuous and discrete Fourier transforms
FYSA21 Mathematical Tools in Science The continuous and discrete Fourier transforms Lennart Lindegren Lund Observatory (Department of Astronomy, Lund University) 1 The continuous Fourier transform 1.1
Membering T M : A Conference Call Service with Speaker-Independent Name Dialing on AIN
PAGE 30 Membering T M : A Conference Call Service with Speaker-Independent Name Dialing on AIN Sung-Joon Park, Kyung-Ae Jang, Jae-In Kim, Myoung-Wan Koo, Chu-Shik Jhon Service Development Laboratory, KT,
English Phonetics: Consonants (i)
1 English Phonetics: Consonants (i) 1.1 Airstream and Articulation Speech sounds are made by modifying an airstream. The airstream we will be concerned with in this book involves the passage of air from
A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation
A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation S.VENKATA RAMANA ¹, S. NARAYANA REDDY ² M.Tech student, Department of ECE, SVU college of Engineering, Tirupati, 517502,
This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.
This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore. Title Transcription of polyphonic signals using fast filter bank( Accepted version ) Author(s) Foo, Say Wei;
Online Diarization of Telephone Conversations
Odyssey 2 The Speaker and Language Recognition Workshop 28 June July 2, Brno, Czech Republic Online Diarization of Telephone Conversations Oshry Ben-Harush, Itshak Lapidot, Hugo Guterman Department of
A TOOL FOR TEACHING LINEAR PREDICTIVE CODING
A TOOL FOR TEACHING LINEAR PREDICTIVE CODING Branislav Gerazov 1, Venceslav Kafedziski 2, Goce Shutinoski 1 1) Department of Electronics, 2) Department of Telecommunications Faculty of Electrical Engineering
Lecture 1-6: Noise and Filters
Lecture 1-6: Noise and Filters Overview 1. Periodic and Aperiodic Signals Review: by periodic signals, we mean signals that have a waveform shape that repeats. The time taken for the waveform to repeat
Broadband Networks. Prof. Dr. Abhay Karandikar. Electrical Engineering Department. Indian Institute of Technology, Bombay. Lecture - 29.
Broadband Networks Prof. Dr. Abhay Karandikar Electrical Engineering Department Indian Institute of Technology, Bombay Lecture - 29 Voice over IP So, today we will discuss about voice over IP and internet
Unified Lecture # 4 Vectors
Fall 2005 Unified Lecture # 4 Vectors These notes were written by J. Peraire as a review of vectors for Dynamics 16.07. They have been adapted for Unified Engineering by R. Radovitzky. References [1] Feynmann,
Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition
, Lisbon Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition Wolfgang Macherey Lars Haferkamp Ralf Schlüter Hermann Ney Human Language Technology
Auto-Tuning Using Fourier Coefficients
Auto-Tuning Using Fourier Coefficients Math 56 Tom Whalen May 20, 2013 The Fourier transform is an integral part of signal processing of any kind. To be able to analyze an input signal as a superposition
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Spot me if you can: Uncovering spoken phrases in encrypted VoIP conversations
Spot me if you can: Uncovering spoken phrases in encrypted VoIP conversations C. Wright, L. Ballard, S. Coull, F. Monrose, G. Masson Talk held by Goran Doychev Selected Topics in Information Security and
MFCC-Based Voice Recognition System for Home Automation Using Dynamic Programming
International Journal of Science and Research (IJSR) MFCC-Based Voice Recognition System for Home Automation Using Dynamic Programming Sandeep Joshi1, Sneha Nagar2 1 PG Student, Embedded Systems, Oriental
Probability and Random Variables. Generation of random variables (r.v.)
Probability and Random Variables Method for generating random variables with a specified probability distribution function. Gaussian And Markov Processes Characterization of Stationary Random Process Linearly
Automatic Detection of Emergency Vehicles for Hearing Impaired Drivers
Automatic Detection of Emergency Vehicles for Hearing Impaired Drivers Sung-won ark and Jose Trevino Texas A&M University-Kingsville, EE/CS Department, MSC 92, Kingsville, TX 78363 TEL (36) 593-2638, FAX
TTT4110 Information and Signal Theory Solution to exam
Norwegian University of Science and Technology Department of Electronics and Telecommunications TTT4 Information and Signal Theory Solution to exam Problem I (a The frequency response is found by taking
How can a speech-language pathologist assess velopharyngeal function without instrumentation?
Clinical Skills for Assessing Velopharyngeal Function by John E. Riski, Ph.D. Speech Pathology at Children's at Scottish Rite http://www.choa.org/ourservices/craniofacial/programs/speech/speechpathology4.asp
AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS
AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS PIERRE LANCHANTIN, ANDREW C. MORRIS, XAVIER RODET, CHRISTOPHE VEAUX Very high quality text-to-speech synthesis can be achieved by unit selection
L3: Organization of speech sounds
L3: Organization of speech sounds Phonemes, phones, and allophones Taxonomies of phoneme classes Articulatory phonetics Acoustic phonetics Speech perception Prosody Introduction to Speech Processing Ricardo
Tutorial about the VQR (Voice Quality Restoration) technology
Tutorial about the VQR (Voice Quality Restoration) technology Ing Oscar Bonello, Solidyne Fellow Audio Engineering Society, USA INTRODUCTION Telephone communications are the most widespread form of transport
Lab 1. The Fourier Transform
Lab 1. The Fourier Transform Introduction In the Communication Labs you will be given the opportunity to apply the theory learned in Communication Systems. Since this is your first time to work in the
Doppler. Doppler. Doppler shift. Doppler Frequency. Doppler shift. Doppler shift. Chapter 19
Doppler Doppler Chapter 19 A moving train with a trumpet player holding the same tone for a very long time travels from your left to your right. The tone changes relative the motion of you (receiver) and
SOFTWARE FOR GENERATION OF SPECTRUM COMPATIBLE TIME HISTORY
3 th World Conference on Earthquake Engineering Vancouver, B.C., Canada August -6, 24 Paper No. 296 SOFTWARE FOR GENERATION OF SPECTRUM COMPATIBLE TIME HISTORY ASHOK KUMAR SUMMARY One of the important
Algorithms for Interference Sensing in Optical CDMA Networks
Algorithms for Interference Sensing in Optical CDMA Networks Purushotham Kamath, Joseph D. Touch and Joseph A. Bannister {pkamath, touch, joseph}@isi.edu Information Sciences Institute, University of Southern
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
SPEECH AUDIOMETRY. @ Biswajeet Sarangi, B.Sc.(Audiology & speech Language pathology)
1 SPEECH AUDIOMETRY Pure tone Audiometry provides only a partial picture of the patient s auditory sensitivity. Because it doesn t give any information about it s ability to hear and understand speech.
Workshop Perceptual Effects of Filtering and Masking Introduction to Filtering and Masking
Workshop Perceptual Effects of Filtering and Masking Introduction to Filtering and Masking The perception and correct identification of speech sounds as phonemes depends on the listener extracting various
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
Speech Analysis for Automatic Speech Recognition
Speech Analysis for Automatic Speech Recognition Noelia Alcaraz Meseguer Master of Science in Electronics Submission date: July 2009 Supervisor: Torbjørn Svendsen, IET Norwegian University of Science and
Speech Recognition System of Arabic Alphabet Based on a Telephony Arabic Corpus
Speech Recognition System of Arabic Alphabet Based on a Telephony Arabic Corpus Yousef Ajami Alotaibi 1, Mansour Alghamdi 2, and Fahad Alotaiby 3 1 Computer Engineering Department, King Saud University,
Automatic Speech Recognition
Automatic Speech Recognition Samudravijaya K Tata Institute of Fundamental Research, Homi Bhabha Road, Mumbai 400005 Email: [email protected] 1 Introduction Information processing machines have become
Speech recognition technology for mobile phones
Speech recognition technology for mobile phones Stefan Dobler Following the introduction of mobile phones using voice commands, speech recognition is becoming standard on mobile handsets. Features such
Predict the Popularity of YouTube Videos Using Early View Data
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
RANDOM VIBRATION AN OVERVIEW by Barry Controls, Hopkinton, MA
RANDOM VIBRATION AN OVERVIEW by Barry Controls, Hopkinton, MA ABSTRACT Random vibration is becoming increasingly recognized as the most realistic method of simulating the dynamic environment of military
Annotated bibliographies for presentations in MUMT 611, Winter 2006
Stephen Sinclair Music Technology Area, McGill University. Montreal, Canada Annotated bibliographies for presentations in MUMT 611, Winter 2006 Presentation 4: Musical Genre Similarity Aucouturier, J.-J.
A comprehensive survey on various ETC techniques for secure Data transmission
A comprehensive survey on various ETC techniques for secure Data transmission Shaikh Nasreen 1, Prof. Suchita Wankhade 2 1, 2 Department of Computer Engineering 1, 2 Trinity College of Engineering and
Marathi Interactive Voice Response System (IVRS) using MFCC and DTW
Marathi Interactive Voice Response System (IVRS) using MFCC and DTW Manasi Ram Baheti Department of CSIT, Dr.B.A.M. University, Aurangabad, (M.S.), India Bharti W. Gawali Department of CSIT, Dr.B.A.M.University,
Advanced Speech-Audio Processing in Mobile Phones and Hearing Aids
Advanced Speech-Audio Processing in Mobile Phones and Hearing Aids Synergies and Distinctions Peter Vary RWTH Aachen University Institute of Communication Systems WASPAA, October 23, 2013 Mohonk Mountain
An Introduction to Neural Networks
An Introduction to Vincent Cheung Kevin Cannons Signal & Data Compression Laboratory Electrical & Computer Engineering University of Manitoba Winnipeg, Manitoba, Canada Advisor: Dr. W. Kinsner May 27,
Audio Scene Analysis as a Control System for Hearing Aids
Audio Scene Analysis as a Control System for Hearing Aids Marie Roch [email protected] Tong Huang [email protected] Jing Liu jliu [email protected] San Diego State University 5500 Campanile Dr San Diego,
Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: [email protected] Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
MODELING DYNAMIC PATTERNS FOR EMOTIONAL CONTENT IN MUSIC
12th International Society for Music Information Retrieval Conference (ISMIR 2011) MODELING DYNAMIC PATTERNS FOR EMOTIONAL CONTENT IN MUSIC Yonatan Vaizman Edmond & Lily Safra Center for Brain Sciences,
CCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York
BME I5100: Biomedical Signal Processing Linear Discrimination Lucas C. Parra Biomedical Engineering Department CCNY 1 Schedule Week 1: Introduction Linear, stationary, normal - the stuff biology is not
Section 5.0 : Horn Physics. By Martin J. King, 6/29/08 Copyright 2008 by Martin J. King. All Rights Reserved.
Section 5. : Horn Physics Section 5. : Horn Physics By Martin J. King, 6/29/8 Copyright 28 by Martin J. King. All Rights Reserved. Before discussing the design of a horn loaded loudspeaker system, it is
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
Implementing an In-Service, Non- Intrusive Measurement Device in Telecommunication Networks Using the TMS320C31
Disclaimer: This document was part of the First European DSP Education and Research Conference. It may have been written by someone whose native language is not English. TI assumes no liability for the
Music Genre Classification
Music Genre Classification Michael Haggblade Yang Hong Kenny Kao 1 Introduction Music classification is an interesting problem with many applications, from Drinkify (a program that generates cocktails
Vector Quantization and Clustering
Vector Quantization and Clustering Introduction K-means clustering Clustering issues Hierarchical clustering Divisive (top-down) clustering Agglomerative (bottom-up) clustering Applications to speech recognition
Measuring Line Edge Roughness: Fluctuations in Uncertainty
Tutor6.doc: Version 5/6/08 T h e L i t h o g r a p h y E x p e r t (August 008) Measuring Line Edge Roughness: Fluctuations in Uncertainty Line edge roughness () is the deviation of a feature edge (as
Articulatory Phonetics. and the International Phonetic Alphabet. Readings and Other Materials. Review. IPA: The Vowels. Practice
Supplementary Readings Supplementary Readings Handouts Online Tutorials The following readings have been posted to the Moodle course site: Contemporary Linguistics: Chapter 2 (pp. 34-40) Handouts for This
Towards better accuracy for Spam predictions
Towards better accuracy for Spam predictions Chengyan Zhao Department of Computer Science University of Toronto Toronto, Ontario, Canada M5S 2E4 [email protected] Abstract Spam identification is crucial
1 Prior Probability and Posterior Probability
Math 541: Statistical Theory II Bayesian Approach to Parameter Estimation Lecturer: Songfeng Zheng 1 Prior Probability and Posterior Probability Consider now a problem of statistical inference in which
PeakVue Analysis for Antifriction Bearing Fault Detection
August 2011 PeakVue Analysis for Antifriction Bearing Fault Detection Peak values (PeakVue) are observed over sequential discrete time intervals, captured, and analyzed. The analyses are the (a) peak values
Sound Pressure Measurement
Objectives: Sound Pressure Measurement 1. Become familiar with hardware and techniques to measure sound pressure 2. Measure the sound level of various sizes of fan modules 3. Calculate the signal-to-noise
