Recent advances in Digital Music Processing and Indexing
|
|
|
- Randell Johnson
- 10 years ago
- Views:
Transcription
1 Recent advances in Digital Music Processing and Indexing Acoustics 08 warm-up TELECOM ParisTech Gaël RICHARD Telecom ParisTech (ENST)
2 Content Introduction and Applications Components of an Audio indexing system Architecture Features extraction and selection Classification : the example of automatic musical instrument recognition Signal Decomposition and source separation An alternative to Bag of frames approaches Non-Negative Matrix decomposition Main melody estimation and extraction Other audio indexing examples Drum Processing :on combining tempo, source separation and transcription Drum loop Retrieval: on combining Speech recognition and transcription Cross-modal Retrieval: On combining visual and audio correlations for audio-based music video retrieval (or video-based audio retrieval). (The evaluation problem) Conclusion
3 Introduction Audio is now widely available in digital format and in huge databases New means of interaction with audio information are possible and desirable Strong needs for efficient audio indexing techniques to automatically obtain a detailed and meaningful symbolic representation
4 Introduction For music signals? Music signal + emotions, playing style, performers,
5 Search by content..?
6 Applications: for content providers Ease composition Search/Retrieval of sound samples or drum loops in large databases of sounds (drum loops retrieval) Content-aware Musical edition software Automatic musical analysis for expressive music synthesis Musical Oracle : predict which kind of public will like a given piece. Hit predictor Search/Retrieval of video scenes from Audio (use of multimodal correlations) Drum loops retrieval reference: O. Gillet and, «Drum loops retrieval from spoken queries», Journal of Intelligent Information Systems, 24:2/3, pp , Springer Science, 2005
7 Applications: for broadcasters Ease radio program set up Navigation interfaces in large data collections Automatic play lists (podcast, ) Mixing/Remixing/DJing : Tempo, rhythm, texture synchronization (Tempo) Identify what is broadcasted Plagiarism detection. Broadcast monitoring. Tempo extraction reference: M. Alonso, and B. David, Accurate tempo estimation based on harmonic+noise decomposition, EURASIP Journal on Advances in Signal Processing, vol. 2007, Article ID 82795, 14 pages, 2007.
8 Applications: for consumers Increase listening capabilities Automatic analysis of user taste for music recommendation. Automatic play list by emotion, genre. Structured navigation in music: «skip the intro, «replay the chorus»,.) Personalized listening (Remixing) Allow new usages Search by content (Query by humming) Smart Karaoke Active listening. Drum Extraction and Remixing: O. Gillet,. Transcription and separation of drum signals from polyphonic music. in IEEE Trans. on Audio, Speech and Language Proc., Volume 16, N 3, March 2008 Page(s):
9 Classification systems Several problems, a similar approach Automatic musical genre recognition Automatic music instruments recognition. Sound samples classification. Sound track labeling (speech, music, special effects etc ). Automatically generated Play list Hit predictor...
10 Traditional Bag-of-Frames approach Preprocessing
11 Features extraction A need for a compact representation of the audio space using descriptors: Temporal parameters : (Zero crossing rate, envelope, amplitude modulation (4 Hz, or Hz), crest factor, signal impulsivity, tempo, Spectral or Cepstral parameters: (MFCC, LPCC, Warped LPCC, spectral centroïd, spectral slope, high order spectral moments, filter banks, harmonic to noise ratio, ) Perceptual parameters: (Relative specific loudness, sharpness and spread,.) Most of these features described in G. Peeters. A large set of audio features for sound description (similarity and classification) in the cuidado project. Technical report, IRCAM (2004) S. Essid. Classification automatique des signaux audio-fréquences: reconnaissance des instruments de musique, Ph.D. thesis, Univ. Paris 6 (in French).
12 Feature extraction «Harmonic» + noise decomposition Example from Subspace decomposition From R. Badeau
13 Classification methods for automatic musical instrument recognition Different strategies were proposed Early studies based on K-Nearest Neighbors (K-NN) on isolated notes and later on solos Current trend: to use more sophisticated modelling approaches such as Discriminant analysis, neural networks, Gaussian Mixtures Models (GMM) and Support Vector Machines See for a recent review: P. Herrera, A. Klapuri, M. Davy. Chap.6 Automatic Classification of Pitched Musical Instrument Sounds. in Signal Processing methods for Music transcription, Edited by A. Klapuri and M. Davy, Springer (2006)
14 Automatic instrument recognition in polyphonic signals From isolated notes to polyphonic signals : a problem of increased complexity Isolated notes solos polyphony In fact, in most cases the methods designed for the monophonic case will not work : Features extraction is non-linear The additivity of the sources cannot be used
15 Automatic instrument recognition in polyphonic signals Some directions: Ignore the features perturbed by other sources (missing features theory) (see Eggink, 2003) Apply a source separation approach before recognising the different instruments (see Gillet 2008 for percussive instruments) Use specific models such as pitch dependent instrument models (Kitahara, 2005) Learn a model for each combination of instruments with an automatically induced hierarchical taxonomy (Essid, 2006)
16 An alternative to bag-of-frames approaches Sparse atomic decomposition of polyphonic music signals Principle: Derive an atomic decomposition of the audio signal. In other words, the signal is approximated as a linear combination of atoms from a fixed dictionary:
17 Sparse atomic decomposition of polyphonic music signals Atoms used: harmonic chirp Gabor atoms (resp ) vector of partials amplitude (resp. phases) scale parameter time localization (resp ) fundamental frequency and chirp rate (from P. Leveau, Décompositions parcimonieuses structurées: Application à la représentation objet de la musique, Ph.D. thesis, Univ. Paris 6, 2007)
18 Sparse atomic decomposition of polyphonic music signals The atomic decomposition is obtained by means of (for example) matching pursuit: First, the most prominent atom (i.e. the most correlated with the signal) is extracted and subtracted from the original signal. Iterate the procedure until a predefined number of atoms have been extracted or until a pre-defined SNR of the representation is reached.
19 Sparse atomic decomposition of polyphonic music signals For musical instrument recognition Use Instrument specific atoms Obtain representatives atoms for each instrument and for each fundamental frequency Use a decomposition algorithm such as matching pursuit Instrument-specific harmonic Molecule can be obtained as a group of successive atoms Démo For more details, see P. Leveau, E. Vincent,, L. Daudet, Instrument-specific harmonic atoms for mid-level music representation, in IEEE Trans. on ASLP, Volume 16, N 1 Jan Page(s):
20 Signal Decomposition and source separation for music transcription Non Negative Matrix factorization Spectrogram W(basis) * H(activation) See for example N. Bertin, R. Badeau,, Blind Signal Decompositions for Automatic transcription of polyphonic music: NMF and K-SVD on the Benchmark, in Proc. of ICASSP 07
21 NMF Towards improved transcription For example use of harmonicity constraints Classic NMF Harmonic NMF with Basis spectra Amplitude coefficients (model the spectral envelope) Narrowband spectra (represent adjacent partials at harmonic frequencies)
22 NMF Towards improved transcription An example of transcription Original Transcription More details in: E. Vincent, N. Bertin and R. Badeau Harmonic and inharmonic nonnegative matrix factorization for polyphonic pitch transcription,, in Proc of ICASSP 08. See also for another approach: V. Emiya, R. Badeau, B. David, Automatic transcription of piano music based on HMM tracking of jointly-estimated pitches, Proc. of EUSIPCO 08 and numerous piano transcription examples:
23 Main melody (Singing voice) extraction Combine NMF (or GMM) and a production model Original spectrum Singing Voice Background Music Background music model : instantaneous mixtures of R Gaussian sources
24 Main melody (Singing voice) extraction Singing voice Model
25 Main melody extraction
26 Main melody (Singing voice) extraction Source Separation (by Wiener Filtering) Two examples Signals Original (mp3) Main voice Music Opera (fem. Voice) Jazz (take 5) More examples at More details in J-L. Durrieu,, B. David, Singer Melody Extraction in Polyphonic signals using source spearation methods, in Proc of ICASSP 08.
27 Some other examples Drum Processing On Combining tempo, source separation and transcription for drum processing Drum loop Retrieval On Combining Speech recognition and transcription for drum loop retrieval Cross-modal Retrieval On combining visual and audio correlations for audiobased music video retrieval (or video-based audio retrieval).
28 Combining tempo, source separation and transcription for drum processing Reference: O. Gillet,. Transcription and separation of drum signals from polyphonic music. in IEEE Trans. on ASLP, Volume 16, N 3, March 2008 Page( s):
29 Combining Speech recognition and transcription for drum loop retrieval Drum loop retrieval from spoken queries An automatic drum loop transcriber A simple HMM based speech recognition engine A retrieval engine that search for the best drum loops based on both textual transcription
30 Combining visual and audio correlations for music video retrieval Music similarity search (or genre classification) For video-based audio retrieval Demo Reference O. Gillet and, «On the Correlation of Automatic Audio and Visual Segmentations of Music Videos» I EEE Trans. on CSVT, 2007
31 The evaluation problem
32 The evaluation problem in Music information indexing and retrieval The domain does not have the historic background of other related domains such as Speech recognition for example. There is a lack of common and public databases and of common protocols for the evaluation Difficulty of obtaining signal annotation under the form of «metadata» (the use of MIDI signals can only partly resolve the problem since MIDI signals are far less complex to process than real audio signals) A big step forward in evaluation: the MIREX experience
33 Evaluation: some results (from MIREX
34 Evaluation: some conclusions A overall percentage of correct recognition has some meaning if: The database is known and well described The evaluation protocol and metrics are well described Comparative evaluations are very important But.evaluations (such as Mirex) are only an instantaneous picture of current algorithms on a given task on a given database with a given protocol
35 Conclusions There is a great interest in exploiting separation and transcription jointly (and this for both individual tasks) Audio indexing technology.is rapidly progressing.is supported by numerous applications still has some limitations especially for the complex case of polyphonic music. still suffers from the lack of common corpora and protocols for rigourous evaluation
36 .Thank you for you attention Contact info and publications:
Automatic Transcription: An Enabling Technology for Music Analysis
Automatic Transcription: An Enabling Technology for Music Analysis Simon Dixon [email protected] Centre for Digital Music School of Electronic Engineering and Computer Science Queen Mary University
Separation and Classification of Harmonic Sounds for Singing Voice Detection
Separation and Classification of Harmonic Sounds for Singing Voice Detection Martín Rocamora and Alvaro Pardo Institute of Electrical Engineering - School of Engineering Universidad de la República, Uruguay
Annotated bibliographies for presentations in MUMT 611, Winter 2006
Stephen Sinclair Music Technology Area, McGill University. Montreal, Canada Annotated bibliographies for presentations in MUMT 611, Winter 2006 Presentation 4: Musical Genre Similarity Aucouturier, J.-J.
WHEN listening to music, humans experience the sound
116 IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 16, NO. 1, JANUARY 2008 Instrument-Specific Harmonic Atoms for Mid-Level Music Representation Pierre Leveau, Emmanuel Vincent, Member,
MUSICAL INSTRUMENT FAMILY CLASSIFICATION
MUSICAL INSTRUMENT FAMILY CLASSIFICATION Ricardo A. Garcia Media Lab, Massachusetts Institute of Technology 0 Ames Street Room E5-40, Cambridge, MA 039 USA PH: 67-53-0 FAX: 67-58-664 e-mail: rago @ media.
AUTOMATIC TRANSCRIPTION OF PIANO MUSIC BASED ON HMM TRACKING OF JOINTLY-ESTIMATED PITCHES. Valentin Emiya, Roland Badeau, Bertrand David
AUTOMATIC TRANSCRIPTION OF PIANO MUSIC BASED ON HMM TRACKING OF JOINTLY-ESTIMATED PITCHES Valentin Emiya, Roland Badeau, Bertrand David TELECOM ParisTech (ENST, CNRS LTCI 46, rue Barrault, 75634 Paris
Audio-visual vibraphone transcription in real time
Audio-visual vibraphone transcription in real time Tiago F. Tavares #1, Gabrielle Odowichuck 2, Sonmaz Zehtabi #3, George Tzanetakis #4 # Department of Computer Science, University of Victoria Victoria,
Speech Signal Processing: An Overview
Speech Signal Processing: An Overview S. R. M. Prasanna Department of Electronics and Electrical Engineering Indian Institute of Technology Guwahati December, 2012 Prasanna (EMST Lab, EEE, IITG) Speech
FAST MIR IN A SPARSE TRANSFORM DOMAIN
ISMIR 28 Session 4c Automatic Music Analysis and Transcription FAST MIR IN A SPARSE TRANSFORM DOMAIN Emmanuel Ravelli Université Paris 6 [email protected] Gaël Richard TELECOM ParisTech [email protected]
BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION
BLIND SOURCE SEPARATION OF SPEECH AND BACKGROUND MUSIC FOR IMPROVED SPEECH RECOGNITION P. Vanroose Katholieke Universiteit Leuven, div. ESAT/PSI Kasteelpark Arenberg 10, B 3001 Heverlee, Belgium [email protected]
HOG AND SUBBAND POWER DISTRIBUTION IMAGE FEATURES FOR ACOUSTIC SCENE CLASSIFICATION. Victor Bisot, Slim Essid, Gaël Richard
HOG AND SUBBAND POWER DISTRIBUTION IMAGE FEATURES FOR ACOUSTIC SCENE CLASSIFICATION Victor Bisot, Slim Essid, Gaël Richard Institut Mines-Télécom, Télécom ParisTech, CNRS LTCI, 37-39 rue Dareau, 75014
Automatic Evaluation Software for Contact Centre Agents voice Handling Performance
International Journal of Scientific and Research Publications, Volume 5, Issue 1, January 2015 1 Automatic Evaluation Software for Contact Centre Agents voice Handling Performance K.K.A. Nipuni N. Perera,
Habilitation. Bonn University. Information Retrieval. Dec. 2007. PhD students. General Goals. Music Synchronization: Audio-Audio
Perspektivenvorlesung Information Retrieval Music and Motion Bonn University Prof. Dr. Michael Clausen PD Dr. Frank Kurth Dipl.-Inform. Christian Fremerey Dipl.-Inform. David Damm Dipl.-Inform. Sebastian
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Antoine Liutkus Researcher, Inria
Antoine Liutkus Researcher, Inria 18 rue des maréchaux 54000 Nancy, France +33662227080 [email protected] born 1981, French Audio filtering and inverse problems Machine Learning Source coding 2014 today
Establishing the Uniqueness of the Human Voice for Security Applications
Proceedings of Student/Faculty Research Day, CSIS, Pace University, May 7th, 2004 Establishing the Uniqueness of the Human Voice for Security Applications Naresh P. Trilok, Sung-Hyuk Cha, and Charles C.
A Sound Analysis and Synthesis System for Generating an Instrumental Piri Song
, pp.347-354 http://dx.doi.org/10.14257/ijmue.2014.9.8.32 A Sound Analysis and Synthesis System for Generating an Instrumental Piri Song Myeongsu Kang and Jong-Myon Kim School of Electrical Engineering,
Control of affective content in music production
International Symposium on Performance Science ISBN 978-90-9022484-8 The Author 2007, Published by the AEC All rights reserved Control of affective content in music production António Pedro Oliveira and
Ericsson T18s Voice Dialing Simulator
Ericsson T18s Voice Dialing Simulator Mauricio Aracena Kovacevic, Anna Dehlbom, Jakob Ekeberg, Guillaume Gariazzo, Eric Lästh and Vanessa Troncoso Dept. of Signals Sensors and Systems Royal Institute of
Advanced Signal Processing and Digital Noise Reduction
Advanced Signal Processing and Digital Noise Reduction Saeed V. Vaseghi Queen's University of Belfast UK WILEY HTEUBNER A Partnership between John Wiley & Sons and B. G. Teubner Publishers Chichester New
Piano Transcription and Radio Travel Model
Unsupervised Transcription of Piano Music Taylor Berg-Kirkpatrick Jacob Andreas Dan Klein Computer Science Division University of California, Berkeley {tberg,jda,klein}@cs.berkeley.edu Abstract We present
EUROPEAN COMPUTER DRIVING LICENCE. Multimedia Audio Editing. Syllabus
EUROPEAN COMPUTER DRIVING LICENCE Multimedia Audio Editing Syllabus Purpose This document details the syllabus for ECDL Multimedia Module 1 Audio Editing. The syllabus describes, through learning outcomes,
MODELING DYNAMIC PATTERNS FOR EMOTIONAL CONTENT IN MUSIC
12th International Society for Music Information Retrieval Conference (ISMIR 2011) MODELING DYNAMIC PATTERNS FOR EMOTIONAL CONTENT IN MUSIC Yonatan Vaizman Edmond & Lily Safra Center for Brain Sciences,
engin erzin the use of speech processing applications is expected to surge in multimedia-rich scenarios
engin erzin Associate Professor Department of Computer Engineering Ph.D. Bilkent University http://home.ku.edu.tr/ eerzin [email protected] Engin Erzin s research interests include speech processing, multimodal
Music Mood Classification
Music Mood Classification CS 229 Project Report Jose Padial Ashish Goel Introduction The aim of the project was to develop a music mood classifier. There are many categories of mood into which songs may
Onset Detection and Music Transcription for the Irish Tin Whistle
ISSC 24, Belfast, June 3 - July 2 Onset Detection and Music Transcription for the Irish Tin Whistle Mikel Gainza φ, Bob Lawlor* and Eugene Coyle φ φ Digital Media Centre Dublin Institute of Technology
EVALUATION OF A SCORE-INFORMED SOURCE SEPARATION SYSTEM
EVALUATION OF A SCORE-INFORMED SOURCE SEPARATION SYSTEM Joachim Ganseman, Paul Scheunders IBBT - Visielab Department of Physics, University of Antwerp 2000 Antwerp, Belgium Gautham J. Mysore, Jonathan
National Standards for Music Education
National Standards for Music Education 1. Singing, alone and with others, a varied repertoire of music. 2. Performing on instruments, alone and with others, a varied repertoire of music. 3. Improvising
SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA
SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA Nitesh Kumar Chaudhary 1 and Shraddha Srivastav 2 1 Department of Electronics & Communication Engineering, LNMIIT, Jaipur, India 2 Bharti School Of Telecommunication,
Music Genre Classification
Music Genre Classification Michael Haggblade Yang Hong Kenny Kao 1 Introduction Music classification is an interesting problem with many applications, from Drinkify (a program that generates cocktails
AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS
AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS PIERRE LANCHANTIN, ANDREW C. MORRIS, XAVIER RODET, CHRISTOPHE VEAUX Very high quality text-to-speech synthesis can be achieved by unit selection
This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore.
This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore. Title Transcription of polyphonic signals using fast filter bank( Accepted version ) Author(s) Foo, Say Wei;
Developing an Isolated Word Recognition System in MATLAB
MATLAB Digest Developing an Isolated Word Recognition System in MATLAB By Daryl Ning Speech-recognition technology is embedded in voice-activated routing systems at customer call centres, voice dialling
A causal algorithm for beat-tracking
A causal algorithm for beat-tracking Benoit Meudic Ircam - Centre Pompidou 1, place Igor Stravinsky, 75004 Paris, France [email protected] ABSTRACT This paper presents a system which can perform automatic
Audio Content Analysis for Online Audiovisual Data Segmentation and Classification
IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 9, NO. 4, MAY 2001 441 Audio Content Analysis for Online Audiovisual Data Segmentation and Classification Tong Zhang, Member, IEEE, and C.-C. Jay
Emotion Detection from Speech
Emotion Detection from Speech 1. Introduction Although emotion detection from speech is a relatively new field of research, it has many potential applications. In human-computer or human-human interaction
The Computer Music Tutorial
Curtis Roads with John Strawn, Curtis Abbott, John Gordon, and Philip Greenspun The Computer Music Tutorial Corrections and Revisions The MIT Press Cambridge, Massachusetts London, England 2 Corrections
Artificial Neural Network for Speech Recognition
Artificial Neural Network for Speech Recognition Austin Marshall March 3, 2005 2nd Annual Student Research Showcase Overview Presenting an Artificial Neural Network to recognize and classify speech Spoken
Advanced Speech-Audio Processing in Mobile Phones and Hearing Aids
Advanced Speech-Audio Processing in Mobile Phones and Hearing Aids Synergies and Distinctions Peter Vary RWTH Aachen University Institute of Communication Systems WASPAA, October 23, 2013 Mohonk Mountain
MUSIC GLOSSARY. Accompaniment: A vocal or instrumental part that supports or is background for a principal part or parts.
MUSIC GLOSSARY A cappella: Unaccompanied vocal music. Accompaniment: A vocal or instrumental part that supports or is background for a principal part or parts. Alla breve: A tempo marking indicating a
A secure face tracking system
International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 4, Number 10 (2014), pp. 959-964 International Research Publications House http://www. irphouse.com A secure face tracking
Audio Engineering Society. Convention Paper. Presented at the 129th Convention 2010 November 4 7 San Francisco, CA, USA
Audio Engineering Society Convention Paper Presented at the 129th Convention 2010 November 4 7 San Francisco, CA, USA The papers at this Convention have been selected on the basis of a submitted abstract
Human behavior analysis from videos using optical flow
L a b o r a t o i r e I n f o r m a t i q u e F o n d a m e n t a l e d e L i l l e Human behavior analysis from videos using optical flow Yassine Benabbas Directeur de thèse : Chabane Djeraba Multitel
Non-negative Matrix Factorization (NMF) in Semi-supervised Learning Reducing Dimension and Maintaining Meaning
Non-negative Matrix Factorization (NMF) in Semi-supervised Learning Reducing Dimension and Maintaining Meaning SAMSI 10 May 2013 Outline Introduction to NMF Applications Motivations NMF as a middle step
Music Transcription with ISA and HMM
Music Transcription with ISA and HMM Emmanuel Vincent and Xavier Rodet IRCAM, Analysis-Synthesis Group 1, place Igor Stravinsky F-704 PARIS [email protected] Abstract. We propose a new generative
Master s Program in Information Systems
The University of Jordan King Abdullah II School for Information Technology Department of Information Systems Master s Program in Information Systems 2006/2007 Study Plan Master Degree in Information Systems
Curriculum Mapping Electronic Music (L) 4202 1-Semester class (18 weeks)
Curriculum Mapping Electronic Music (L) 4202 1-Semester class (18 weeks) Week Standard Skills Resources Vocabulary Assessments Students sing using computer-assisted instruction and assessment software.
International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014
RESEARCH ARTICLE OPEN ACCESS A Survey of Data Mining: Concepts with Applications and its Future Scope Dr. Zubair Khan 1, Ashish Kumar 2, Sunny Kumar 3 M.Tech Research Scholar 2. Department of Computer
Using Support Vector Machines for Automatic Mood Tracking in Audio Music
Audio Engineering Society Convention Paper Presented at the 130th Convention 2011 May 13 16 London, UK The papers at this Convention have been selected on the basis of a submitted abstract and extended
School Class Monitoring System Based on Audio Signal Processing
C. R. Rashmi 1,,C.P.Shantala 2 andt.r.yashavanth 3 1 Department of CSE, PG Student, CIT, Gubbi, Tumkur, Karnataka, India. 2 Department of CSE, Vice Principal & HOD, CIT, Gubbi, Tumkur, Karnataka, India.
M3039 MPEG 97/ January 1998
INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND ASSOCIATED AUDIO INFORMATION ISO/IEC JTC1/SC29/WG11 M3039
Salisbury Township School District Planned Course of Study - Music Production Salisbury Inspire, Think, Learn, Grow Together!
Topic/Unit: Music Production: Recording and Microphones Suggested Timeline: 1-2 weeks Big Ideas/Enduring Understandings: Students will listen to, analyze, and describe music that has been created using
Carla Simões, [email protected]. Speech Analysis and Transcription Software
Carla Simões, [email protected] Speech Analysis and Transcription Software 1 Overview Methods for Speech Acoustic Analysis Why Speech Acoustic Analysis? Annotation Segmentation Alignment Speech Analysis
Beethoven, Bach und Billionen Bytes
Meinard Müller Beethoven, Bach und Billionen Bytes Automatisierte Analyse von Musik und Klängen Meinard Müller Tutzing-Symposium Oktober 2014 2001 PhD, Bonn University 2002/2003 Postdoc, Keio University,
Semantic Video Annotation by Mining Association Patterns from Visual and Speech Features
Semantic Video Annotation by Mining Association Patterns from and Speech Features Vincent. S. Tseng, Ja-Hwung Su, Jhih-Hong Huang and Chih-Jen Chen Department of Computer Science and Information Engineering
Greenwich Public Schools Electronic Music Curriculum 9-12
Greenwich Public Schools Electronic Music Curriculum 9-12 Overview Electronic Music courses at the high school are elective music courses. The Electronic Music Units of Instruction include four strands
A Pattern Recognition Approach for Melody Track Selection in MIDI Files
A Pattern Recognition Approach for Melody Track Selection in MIDI Files David Rizo, Pedro J. Ponce de León, Carlos Pérez-Sancho, Antonio Pertusa, José M. Iñesta Departamento de Lenguajes y Sistemas Informáticos
Machine Learning with MATLAB David Willingham Application Engineer
Machine Learning with MATLAB David Willingham Application Engineer 2014 The MathWorks, Inc. 1 Goals Overview of machine learning Machine learning models & techniques available in MATLAB Streamlining the
Adaptive Radio. Cognitive Radio
What are Cognitive Radio and Dynamic Spectrum Access SDR can act as a key enabling technology for a variety of other reconfigurable radio equipments commonly discussed in the advanced wireless market 1.
MIRACLE at VideoCLEF 2008: Classification of Multilingual Speech Transcripts
MIRACLE at VideoCLEF 2008: Classification of Multilingual Speech Transcripts Julio Villena-Román 1,3, Sara Lana-Serrano 2,3 1 Universidad Carlos III de Madrid 2 Universidad Politécnica de Madrid 3 DAEDALUS
THE SEMANTIC HIFI PROJECT
THE SEMANTIC HIFI PROJECT Hugues Vinet IRCAM-CNRS STMS 1, place Igor Stravinsky 75004 PARIS FRANCE [email protected] ABSTRACT The SemanticHIFI European project aims at designing and prototyping tomorrow
Monophonic Music Recognition
Monophonic Music Recognition Per Weijnitz Speech Technology 5p [email protected] 5th March 2003 Abstract This report describes an experimental monophonic music recognition system, carried out
QMeter Tools for Quality Measurement in Telecommunication Network
QMeter Tools for Measurement in Telecommunication Network Akram Aburas 1 and Prof. Khalid Al-Mashouq 2 1 Advanced Communications & Electronics Systems, Riyadh, Saudi Arabia [email protected] 2 Electrical
Graduate Co-op Students Information Manual. Department of Computer Science. Faculty of Science. University of Regina
Graduate Co-op Students Information Manual Department of Computer Science Faculty of Science University of Regina 2014 1 Table of Contents 1. Department Description..3 2. Program Requirements and Procedures
Quarterly Progress and Status Report. Measuring inharmonicity through pitch extraction
Dept. for Speech, Music and Hearing Quarterly Progress and Status Report Measuring inharmonicity through pitch extraction Galembo, A. and Askenfelt, A. journal: STL-QPSR volume: 35 number: 1 year: 1994
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
PIANO MUSIC TRANSCRIPTION WITH FAST CONVOLUTIONAL SPARSE CODING
25 IEEE INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING, SEPT. 7 2, 25, BOSTON, USA PIANO MUSIC TRANSCRIPTION WITH FAST CONVOLUTIONAL SPARSE CODING Andrea Cogliati, Zhiyao Duan University
From the concert hall to the library portal
The following is a written transcription of a presentation made at IASA 2006 conference in Mexico city by Marie-Hélène Serra and Rodolphe Bailly. From the concert hall to the library portal Since this
Modeling Affective Content of Music: A Knowledge Base Approach
Modeling Affective Content of Music: A Knowledge Base Approach António Pedro Oliveira, Amílcar Cardoso Centre for Informatics and Systems of the University of Coimbra, Coimbra, Portugal Abstract The work
DIGITAL MUSIC DAY 1 WHAT IS SOUND? ANALOG AND DIGITAL EARLY RECORDING WAX FOR YOUR EARS ROUND BUT FLAT WIRE AND TAPE PURE SOUND
DIGITAL MUSIC DAY 1 WHAT IS SOUND? 1. Making a sound playing a musical instrument, moves the surrounding the instrument. 2. Making a sound sends out air which hit your ears. 3. Waves of changing air pressure
IT services for analyses of various data samples
IT services for analyses of various data samples Ján Paralič, František Babič, Martin Sarnovský, Peter Butka, Cecília Havrilová, Miroslava Muchová, Michal Puheim, Martin Mikula, Gabriel Tutoky Technical
LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE. [email protected]
LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE 1 S.Manikandan, 2 S.Abirami, 2 R.Indumathi, 2 R.Nandhini, 2 T.Nanthini 1 Assistant Professor, VSA group of institution, Salem. 2 BE(ECE), VSA
Electronic music. live electronic music is where technology is used to make, change, or play sounds during live performance (usually on laptops).
Electronic music Electronic music is music in which electronic technology, now mostly computer-based, is used to access, make, explore and configure sound materials, and which is heard through loudspeakers.
TEMPO AND BEAT ESTIMATION OF MUSICAL SIGNALS
TEMPO AND BEAT ESTIMATION OF MUSICAL SIGNALS Miguel Alonso, Bertrand David, Gaël Richard ENST-GET, Département TSI 46, rue Barrault, Paris 75634 cedex 3, France {malonso,bedavid,grichard}@tsi.enst.fr ABSTRACT
Lecture 1-10: Spectrograms
Lecture 1-10: Spectrograms Overview 1. Spectra of dynamic signals: like many real world signals, speech changes in quality with time. But so far the only spectral analysis we have performed has assumed
HIDDEN MARKOV MODELS FOR SPECTRAL SIMILARITY OF SONGS. Arthur Flexer, Elias Pampalk, Gerhard Widmer
Proc. of the th Int. Conference on Digital Audio Effects (DAFx 5), Madrid, Spain, September 2-22, 25 HIDDEN MARKOV MODELS FOR SPECTRAL SIMILARITY OF SONGS Arthur Flexer, Elias Pampalk, Gerhard Widmer The
NAVIGATING SCIENTIFIC LITERATURE A HOLISTIC PERSPECTIVE. Venu Govindaraju
NAVIGATING SCIENTIFIC LITERATURE A HOLISTIC PERSPECTIVE Venu Govindaraju BIOMETRICS DOCUMENT ANALYSIS PATTERN RECOGNITION 8/24/2015 ICDAR- 2015 2 Towards a Globally Optimal Approach for Learning Deep Unsupervised
RESEARCH IN MUSIC EDUCATION AND SOUND ARTS (Completed in, and/or about New Zealand)
Te Puna Puoru National Centre for Research in Music Education and Sound Arts (MERC) invites you to contribute to the national database of research in music education and sound arts. Fill in the template
Multimedia Databases. Wolf-Tilo Balke Philipp Wille Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.
Multimedia Databases Wolf-Tilo Balke Philipp Wille Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 0 Organizational Issues Lecture 21.10.2014 03.02.2015
An Introduction to Data Mining. Big Data World. Related Fields and Disciplines. What is Data Mining? 2/12/2015
An Introduction to Data Mining for Wind Power Management Spring 2015 Big Data World Every minute: Google receives over 4 million search queries Facebook users share almost 2.5 million pieces of content
DYNAMIC CHORD ANALYSIS FOR SYMBOLIC MUSIC
DYNAMIC CHORD ANALYSIS FOR SYMBOLIC MUSIC Thomas Rocher, Matthias Robine, Pierre Hanna, Robert Strandh LaBRI University of Bordeaux 1 351 avenue de la Libration F 33405 Talence, France [email protected]
Hardware Implementation of Probabilistic State Machine for Word Recognition
IJECT Vo l. 4, Is s u e Sp l - 5, Ju l y - Se p t 2013 ISSN : 2230-7109 (Online) ISSN : 2230-9543 (Print) Hardware Implementation of Probabilistic State Machine for Word Recognition 1 Soorya Asokan, 2
