DECODA: a call-center human-human spoken conversation corpus
|
|
|
- Pierce Doyle
- 10 years ago
- Views:
Transcription
1 DECODA: a call-center human-human spoken conversation corpus F. Bechet 1, B. Maza 2, N. Bigouroux 3, T. Bazillon 1, M. El-Bèze 2, R. De Mori 2, E. Arbillot 4 1 Aix Marseille Univ, LIF-CNRS, Marseille, France [email protected] [email protected] 2 Universite d Avignon, Avignon, France [email protected] [email protected] [email protected] 3 Sonear, Paris, France [email protected] 4 RATP, Paris, France [email protected] Abstract The goal of the DECODA project is to reduce the development cost of Speech Analytics systems by reducing the need for manual annotation. This project aims to propose robust speech data mining tools in the framework of call-center monitoring and evaluation, by means of weakly supervised methods. The applicative framework of the project is the call-center of the RATP (Paris public transport authority). This project tackles two very important open issues in the development of speech mining methods from spontaneous speech recorded in call-centers : robustness (how to extract relevant information from very noisy and spontaneous speech messages) and weak supervision (how to reduce the annotation effort needed to train and adapt recognition and classification models). This paper describes the DECODA corpus collected at the RATP during the project. We present the different annotation levels performed on the corpus, the methods used to obtain them, as well as some evaluation of the quality of the annotations produced. Keywords: Speech Analytics, Conversational Speech, Syntactic and Semantic annotations 1. Introduction Speech Analytics is a new term used to describe the extraction of information from speech data. Three main research domains are involved: Automatic Speech Recognition (ASR), Natural Language Processing (NLP) and Data Mining. The first level in the Speech Analytics process is to translate a speech signal into several sequences of symbols. These symbols can describe both the lexical hypotheses about the linguistic content of a speech segment and the acoustic and prosodic properties of it such as: environmental noise, quality of speech, speaker, prosodic cues, etc. This first level is made by the Signal Processing and Automatic Speech Recognition processes (ASR). The lexical hypotheses produced by the ASR modules are processed by the NLP modules. Then the Data Mining processed can extract from a large corpus some information about the behaviors of the callers in the context of a call-center. The multiplication of call-centers and the low storing cost of audio data have made available the recording of very large audio database containing conversations between operators and customers. From the companies point of view, these call-centers are a strategic interface toward their customers. Therefore the need for automatic tools allowing to monitor such services and mine the dialogs recorded is very high, as pointed out by the development of industrial products, mainly by American companies, in the field of Speech Analytics: Nuance, Verint, CallMiner, BBN-Avoke, Nexidia, Autonomy etalk. These products can provide two kinds of services: punctual analysis of large dialog corpora for data mining purposes, like detecting a problem in the callcenter behavior, or extracting knowledge about the call-center performance; periodic analysis, or monitoring of the call-center by a day-by-day analysis of the call-center dialog logs. An example of a speech analytics process is displayed in figure 1, where a conversation between an agent and a caller is analyzed according to several dimensions: from the segmentation of the dialogs, the extraction of concepts and meanings to the summarization process and the behavioral analysis. All the products need the manual annotation of large amount of speech data in order to train and adapt the recognition and classification models. This task has to be done regularly in order to adapt the models to the new behaviors of a call-center, leading to an increase in the cost development of Speech Analytics tools. The goal of the DECODA 1 project is to reduce the development cost of Speech Analytics systems by leveraging the manual annotation effort needed. This project develops robust speech data mining tools in the framework of callcenter monitoring and evaluation, by means of weakly supervised methods. This project tackles two very important open issues in the development of speech mining methods from spontaneous speech recorded in call-centers: robustness (how to extract relevant information from very noisy and spontaneous speech messages) and weak supervision (how to reduce the annotation effort needed to train and adapt recognition and classification models). The applicative framework of the project is the call-center of the RATP (Paris public transport company). This paper describes the DECODA corpus collected at the RATP during the project. We present the different annotation levels performed on the corpus, the methods used to
2 - Good morning, how may I help you? - Hello. I am calling regarding my internet connection. - What's your phone number please? - It's I see, you have a connection problem, right? - Yes, it's the fourth time I call about that. - Okay, can you check that the light is... Opening Information Verification Conflict Situtation Closing - Goodbye. Problem Resolution - No, that's fine, goodbye. - Do you need anything else? - Oh great, thanks. - Hmm... - Alright, it doesn't work anymore. We will send you a new one shortly. - Sir, please calm down. Let me remotely check your modem. - I already did all the checks last time! Can't you just send a repairman? Segment Conflict about procedure Emotion Meaning procedure[state=completed,time=past] query[obj=intervention,loc=home] Relevance Dialog is problematic Behaviour Analysis Answer complex questions about conversations: - Was the operator empathic to customer's concerns? - Was the customer satistified? - How does the agent manage the argument? - Did the agent achieve the goal of the call? - Is the agent polite? Figure 1: Example of speech analytic processes on a call-center conversation obtain them, as well as some evaluations about the quality of the annotations produced. 2. Related work: previous corpora of spoken conversation Most of the studies of human-human spoken conversations have been limited in scope due to either size of the corpora or realistic task scenarios with only a few exceptions (Tur and Hakkani-Tür, 2011). Several corpora have been collected in the past and used for conversation analysis. The first example is probably Switchboard (Godfrey et al., 1992), a large corpus of telephone conversations involving about 500 speakers who were given a pre-determined topic to talk about. The research originally focused on speech and speaker recognition rather than speech understanding. More natural and multilingual/accented conversational speech was collected later in the Call-Home and Call-Friend corpora. Goal-orientated human-human dialogues are available in the TRAINS (Allen, 1994) corpus for transportation planning, in the Monroe (Stent, 2000) corpus for disaster handling, and in the MapTask (Anderson et al., 1991) corpus for giving directions on a map. More recently, multi-party human-human conversations (or meetings) and lectures have been considered, for example in the European funded projects AMI (Augmented Multiparty Interaction) and AMIDA (Renals et al., 2007), and, in the USA, the CALO (Cognitive Assistant that Learns and Organizes) project that concentrated on conference-room meetings with small scale in-vitro experiments. A corpus of university lectures, some of them including interactions with the audience, was used in the CHIL (Computers in the Human Interaction Loop) European funded project. These types of corpora inspired research on human-human conversation mostly with specific purposes such as browsing and learning dialogue structures (e.g. dialogue act tagging, topic segmentation, summarization). Another source of human-human conversation corpora is broadcast conversations such as debates or interviews. This kind of data has the advantage of being easily accessible from corpus distribution agencies such as LDC or ELDA through large corpora of broadcast shows. However broadcast conversations are far from being truly spontaneous as many of the speakers are professionals. Telecommunication and call-center companies involved in Customer Relationship Management (CRM), internally record call-center conversational speech data, which are the focus of this paper. Recently some European and national research programs have devoted resources to acquiring corpora through collaboration between service providers and research labs. In the LUNA FP6 project human-machine and humanhuman conversations have been collected in the context of Information Technology (IT) help-desk interactions. In France the CallSurf project (Garnier-Rizet et al., 2008) has collected a large corpus of conversations through an EDF call-center. Part of this corpus has been manually tran- 1344
3 scribed and annotated. 3. The DECODA corpus The main problem with call-center data is that it often contains a large amount of personal data informations, belonging to the clients of the call-center. The conversations collected are very difficult to anonymized, unless large amounts of signal are erased, and therefore the corpus collected can t be distributed toward the scientific community. In the DECODA project we are dealing with the call-center of the Paris transport authority (RATP). This applicative framework is very interesting because it allows us to easily collect large amount of data, from a large range of speakers, with very few personal data. Indeed people hardly introduce themselves while phoning to obtain bus or subway directions, ask for a lost luggage or for information about the traffic. Therefore this kind of data can be anonymized without erasing a lot of signal. The DECODA corpus currently collected within the project has been fully anonymized, manually segmented and transcribed, then annotated at various linguistic levels. The current state of the corpus is made of 1514 dialogs, corresponding to about 74 hours of signal. The average duration of a dialog is about 3 minutes. The distribution of the duration is given in table 1. As we can see most of dialogs are quite short, however 12% of them are longer than 5 minutes. The collect of the first 1514 dialogs of the DECODA corpus has been done during 2 days of traffic, in order to have a good picture of all the possible requests expressed by the callers during a whole day. 4. Speech transcription The speech transcription process involved three steps: the anonymization process; the segmentation process and the transcription process itself. The goal of the anonymization process is to suppress from the audio files all the personal references such as person names, phone numbers or personal addresses. This is a fully manual process, made before any other tasks, and consisting of replacing all these personal references by a beep. Thankfully, as mentioned before, there is very little personal information in the corpus, except in the first sentence where the two speakers identify themselves. The second step is a segmentation process. The whole audio file is first segmented into chapters: opening, problem presentation, problem resolution, closing. Then a speaker segmentation process is performed in order to separate the turns of each speaker. This step is necessary since unfortunately the DECODA corpus is recorded with only one channel mixing the voices of the callers and the operators. Finally each turn is segmented into sentence-like units (called segments), separated within each other by a pause longer than a given threshold. All the sentence-like units have been manually transcribed using the tool Transcriber and the conventions of the ES- TER (Galliano et al., 2009) program. The DECODA corpus contains so far 1514 dialogs, resulting into speakers turns, segments for a total of words after tokenization. The most frequent word in this corpus is the discourse marker euh (hum), indicating the spontaneity of the speech collected. The total vocabulary of the corpus is 8806 words. 5. Semantic annotation At the semantic level each dialog has been manually tagged according to the RATP ontology. The repartition of the 10 top call-types is given in table 2. Calltype % Info Traffic 22.5 Route planning 17.2 Lost and Found 15.9 Registration card 11.4 Timetable 4.6 Ticket 4.5 Specialized calls 4.5 empty 3.6 New registration 3.4 Price info 3.0 Table 2: Repartion of the 10 top calltypes in the DECODA corpus As we can see, 4 call-types represent 67% of the dialogs. The top call-type is Info Traffic which contains requests about delays, roadworks and strikes in the transportation network. This is an interesting category as a large variety of customers behaviors occur, like for example stress and angriness sentiments from people being late because of a problem on the transportation network. 6. Syntactic annotation Five levels of linguistic annotation have been performed on the transcriptions of the DECODA corpus: 1. Disfluencies 2. Part-of-Speech (POS) tags 3. Chunks 4. Named Entities 5. Syntactic dependencies The first level (disfluencies) corresponds to repetitions (e.g. le le), discourse markers (e.g. euh, bien) and false starts (e.g. bonj-). These annotations have been manually performed on the corpus thanks to a WEB-interface allowing to write regular expressions to recognize some word patterns corresponding to a kind of disfluencies. For each regular expression written, the interface displays all the matches found in the corpus in order to check the relevance of the expression. Once validated, all the occurrences matching the expression are tagged with the chosen label. Table 3 displays the amount of disfluencies found in the corpus, according to their types, as well as the most frequent ones. As we can see, discourse markers are by far the most frequent type of disfluencies, occurring in 28% of the speech segments. The Named Entity annotation level has been performed manually with the same WEB-interface. In our corpus, 1345
4 Duration (min.) <= > 10 # of dialogs Table 1: Repartition of the duration disfluency type # occ. % of turns 10 most frequent forms discourse markers % [euh] [hein] [ah] [ben] [voilà] [bon] [hm] [bah] [hm hm] [écoutez] repetitions % [oui oui] [non non] [c est c est] [le le] [de de] [ouais ouais] [je je] [oui oui oui] [non non non] [ça ça] false starts % [s-] [p-] [l-] [m-] [d-] [v-] [c-] [t-] [b-] [n-] Table 3: Distribution of the disfluencies manually annotated on the DECODA corpus most of the named entities correspond to location expressions such as addresses, bus stops and metro stations. The POS and chunk levels of annotation have been first automatically produced by the NLP tool MACAON (Nasr et al., 2011). At the same time a fully manual annotation process is performed on a subset of the corpus, called the GOLD corpus. An iterative process is then applied, consisting in manually correcting errors found in the automatic annotations, retraining the linguistic models of the NLP tools on this corrected corpus, then checking the quality of the adapted models on the fully manual annotations of the GOLD corpus. This process iterates until a certain error rate is reached (Bazillon et al., 2012). Table 4 shows the POS error rate before and after this adaptation process on the GOLD corpus. As we can see the error rate is reduced by almost 60%. POS error rate Baseline After adaptation DECODA GOLD corpus 21.0% 8.5% Table 4: POS error rate before and after the iterative adaptation process performed on the DECODA training corpus The same kind of semi-supervised method has been used to obtain the syntactic dependency annotations. These kinds of annotation are very useful in order to train new approaches to parsing based on dependency structures and discriminative machine learning techniques (Nivre, 2003; McDonald et al., 2005). These approaches are much easier to adapt to speech than context-free grammar approaches since they need less training data and the annotation with syntactic dependencies of speech transcripts is simpler than with syntactic constituents. The annotation process of the DECODA corpus is described in details in (Bazillon et al., 2012). A first evaluation of a statistical dependency parser trained on this annotated corpus and applied to the GOLD corpus has obtained encouraging results. An example of a full annotated segment is displayed in table 5. The corpus can be found either on this semicolon format or in an XML format, similar to the one used in MACAON (Nasr et al., 2011). 7. Conclusion We have described in this paper the acquisition, anonymization, transcription and annotation process of the DECODA corpus collected through the RATP call-center in Paris. The first experiments we carried out have controlled the quality of the annotations produced (Bazillon et al., 2012) and the usefulness of the corpus for spoken language understanding tasks (Maza et al., 2011). We are going now to fully exploit this annotated corpus in order to propose advanced speech analytics systems taking into account all the levels of annotation produced. 8. Acknowledgements This work is supported by the French agency ANR, Project DECODA, contract no 2009-CORD , and the French business clusters Cap Digital and SCS. For more information about the DE- CODA project, please visit the project home-page, 9. References J.F. Allen The trains project: A case study in building a conversational planning agent. Technical report, DTIC Document. A.H. Anderson, M. Bader, E.G. Bard, E. Boyle, G. Doherty, S. Garrod, S. Isard, J. Kowtko, J. McAllister, J. Miller, et al The hcrc map task corpus. Language and speech, 34(4):351. Thierry Bazillon, Melanie Deplano, Frederic Bechet, Alexis Nasr, and Benoit Favre Syntactic annotation of spontaneous speech: application to call-center conversation data. In Proceedings of LREC, Istambul. S. Galliano, G. Gravier, and L. Chaubard The ester 2 evaluation campaign for the rich transcription of french radio broadcasts. In Tenth Annual Conference of the International Speech Communication Association. M. Garnier-Rizet, G. Adda, F. Cailliau, J.L. Gauvain, S. Guillemin-Lanne, L. Lamel, S. Vanni, and C. Waast- Richard Callsurf-automatic transcription, indexing and structuration of call center conversational speech for knowledge extraction and query by content. In Proceedings of LREC. J.J. Godfrey, E.C. Holliman, and J. McDaniel Switchboard: Telephone speech corpus for research and development. In Acoustics, Speech, and Signal Processing, ICASSP-92., 1992 IEEE International Conference on, volume 1, pages IEEE. Benjamin Maza, Marc El-Beze, Georges Linares, and Renato De Mori On the use of linguistic features 1346
5 id word disfluencies NE POS chunk dep. link dep. type 1 oui - - adv B_GADV 0 ROOT 2 je B_REP - cln je I_REP - cln je - - cln B_GN 5 SUJ 5 sais - - v B_GVfn 0 ROOT 6 pas - - advneg I_GVfn 5 MOD 7 euh B_DISM - pres je - - cln B_GN 14 SUJ 9 ne - - clneg B_GVfn 12 MOD 10 m - - clr I_GVfn 12 AFF 11 en - - clo I_GVfn 14 DE_OBJ 12 étais - - v I_GVfn 14 AUX 13 pas - - advneg I_GVfn 12 MOD 14 inquiétée - - vppart I_GVfn 0 ROOT 15 madame - - nc B_GN 0 root 16 c - - cln B_GN 17 SUJ_IMP 17 est - - v B_GVf 0 ROOT 18 École - B_LOC np B_GN 17 ATS 19 Véterinaire - I_LOC np ou - - coo B_coo 18 COORD 21 Rungis - B_LOC np B_GN 20 DEP_COORD Table 5: Example of annotated segment in the DECODA corpus in an automatic system for speech analytics of telephone conversations. In Proceedings of Interspeech 2011, Florence, Italy. Ryan McDonald, Koby Crammer, and Fernando Pereira Online Large-Margin Training of Dependency Parsers. In Association for Computational Linguistics (ACL). Alexis Nasr, Frederic Bechet, Jean-Francois Rey, and Joseph Le Roux Macaon:a linguistic tool suite for processing word lattices. In The 49th Annual Meeting of the Association for Computational Linguistics: demonstration session. J. Nivre An efficient algorithm for projective dependency parsing. In Proceedings of the 8th International Workshop on Parsing Technologies (IWPT. S. Renals, T. Hain, and H. Bourlard Recognition and interpretation of meetings: The ami and amida projects. A.J. Stent The monroe corpus. TR728 and TN99-2, Dept. of Computer Science, University of Rochester. Gokhan Tur and Dilek Hakkani-Tür Human/human conversation understanding. Spoken Language Understanding: Systems for Extracting Semantic Information from Speech, Gokhan Tur and Renato De Mori ed. Wiley. 1347
Syntactic annotation of spontaneous speech: application to call center conversation data
Syntactic annotation of spontaneous speech: application to call center conversation data Frédéric Béchet, Thierry Bazillon, Benoit Favre, Alexis Nasr Aix Marseille Université LIF-CNRS Laboratoire d Informatique
CallAn: A Tool to Analyze Call Center Conversations
CallAn: A Tool to Analyze Call Center Conversations Balamurali AR, Frédéric Béchet And Benoit Favre Abstract Agent Quality Monitoring (QM) of customer calls is critical for call center companies. We present
Call Centre Conversation Summarization: A Pilot Task at Multiling 2015
Call Centre Conversation Summarization: A Pilot Task at Multiling 2015 Benoit Favre 1, Evgeny Stepanov 2, Jérémy Trione 1, Frédéric Béchet 1, Giuseppe Riccardi 2 1 Aix-Marseille University, CNRS, LIF UMR
D2.4: Two trained semantic decoders for the Appointment Scheduling task
D2.4: Two trained semantic decoders for the Appointment Scheduling task James Henderson, François Mairesse, Lonneke van der Plas, Paola Merlo Distribution: Public CLASSiC Computational Learning in Adaptive
Robust Methods for Automatic Transcription and Alignment of Speech Signals
Robust Methods for Automatic Transcription and Alignment of Speech Signals Leif Grönqvist ([email protected]) Course in Speech Recognition January 2. 2004 Contents Contents 1 1 Introduction 2 2 Background
The REPERE Corpus : a multimodal corpus for person recognition
The REPERE Corpus : a multimodal corpus for person recognition Aude Giraudel 1, Matthieu Carré 2, Valérie Mapelli 2 Juliette Kahn 3, Olivier Galibert 3, Ludovic Quintard 3 1 Direction générale de l armement,
Specialty Answering Service. All rights reserved.
0 Contents 1 Introduction... 2 1.1 Types of Dialog Systems... 2 2 Dialog Systems in Contact Centers... 4 2.1 Automated Call Centers... 4 3 History... 3 4 Designing Interactive Dialogs with Structured Data...
CallSurf - Automatic transcription, indexing and structuration of call center conversational speech for knowledge extraction and query by content
CallSurf - Automatic transcription, indexing and structuration of call center conversational speech for knowledge extraction and query by content Martine Garnier-Rizet 1, Gilles Adda 2, Frederik Cailliau
Research Portfolio. Beáta B. Megyesi January 8, 2007
Research Portfolio Beáta B. Megyesi January 8, 2007 Research Activities Research activities focus on mainly four areas: Natural language processing During the last ten years, since I started my academic
Automatic Speech Recognition and Hybrid Machine Translation for High-Quality Closed-Captioning and Subtitling for Video Broadcast
Automatic Speech Recognition and Hybrid Machine Translation for High-Quality Closed-Captioning and Subtitling for Video Broadcast Hassan Sawaf Science Applications International Corporation (SAIC) 7990
The Knowledge Sharing Infrastructure KSI. Steven Krauwer
The Knowledge Sharing Infrastructure KSI Steven Krauwer 1 Why a KSI? Building or using a complex installation requires specialized skills and expertise. CLARIN is no exception. CLARIN is populated with
Speech Transcription
TC-STAR Final Review Meeting Luxembourg, 29 May 2007 Speech Transcription Jean-Luc Gauvain LIMSI TC-STAR Final Review Luxembourg, 29-31 May 2007 1 What Is Speech Recognition? Def: Automatic conversion
Module Catalogue for the Bachelor Program in Computational Linguistics at the University of Heidelberg
Module Catalogue for the Bachelor Program in Computational Linguistics at the University of Heidelberg March 1, 2007 The catalogue is organized into sections of (1) obligatory modules ( Basismodule ) that
Turkish Radiology Dictation System
Turkish Radiology Dictation System Ebru Arısoy, Levent M. Arslan Boaziçi University, Electrical and Electronic Engineering Department, 34342, Bebek, stanbul, Turkey [email protected], [email protected]
Evaluation of speech technologies
CLARA Training course on evaluation of Human Language Technologies Evaluations and Language resources Distribution Agency November 27, 2012 Evaluation of speaker identification Speech technologies Outline
Terminology Extraction from Log Files
Terminology Extraction from Log Files Hassan Saneifar 1,2, Stéphane Bonniol 2, Anne Laurent 1, Pascal Poncelet 1, and Mathieu Roche 1 1 LIRMM - Université Montpellier 2 - CNRS 161 rue Ada, 34392 Montpellier
Accelerating and Evaluation of Syntactic Parsing in Natural Language Question Answering Systems
Accelerating and Evaluation of Syntactic Parsing in Natural Language Question Answering Systems cation systems. For example, NLP could be used in Question Answering (QA) systems to understand users natural
The LENA TM Language Environment Analysis System:
FOUNDATION The LENA TM Language Environment Analysis System: The Interpreted Time Segments (ITS) File Dongxin Xu, Umit Yapanel, Sharmi Gray, & Charles T. Baer LENA Foundation, Boulder, CO LTR-04-2 September
A System for Labeling Self-Repairs in Speech 1
A System for Labeling Self-Repairs in Speech 1 John Bear, John Dowding, Elizabeth Shriberg, Patti Price 1. Introduction This document outlines a system for labeling self-repairs in spontaneous speech.
Proper Name Retrieval from Diachronic Documents for Automatic Speech Transcription using Lexical and Temporal Context
Proper Name Retrieval from Diachronic Documents for Automatic Speech Transcription using Lexical and Temporal Context Irina Illina, Dominique Fohr, Georges Linarès To cite this version: Irina Illina, Dominique
CallMiner Speech Analytics Everything else is just talk. Cliff LaCoursiere SVP Business Development - CallMiner, Inc.
CallMiner Speech Analytics Everything else is just talk. Cliff LaCoursiere SVP Business Development - CallMiner, Inc. Agenda Why speech analytics? How CallMiner speech analytics works Speech analytics
What do Big Data & HAVEn mean? Robert Lejnert HP Autonomy
What do Big Data & HAVEn mean? Robert Lejnert HP Autonomy Much higher Volumes. Processed with more Velocity. With much more Variety. Is Big Data so big? Big Data Smart Data Project HAVEn: Adaptive Intelligence
Terminology Extraction from Log Files
Terminology Extraction from Log Files Hassan Saneifar, Stéphane Bonniol, Anne Laurent, Pascal Poncelet, Mathieu Roche To cite this version: Hassan Saneifar, Stéphane Bonniol, Anne Laurent, Pascal Poncelet,
Semantic annotation of requirements for automatic UML class diagram generation
www.ijcsi.org 259 Semantic annotation of requirements for automatic UML class diagram generation Soumaya Amdouni 1, Wahiba Ben Abdessalem Karaa 2 and Sondes Bouabid 3 1 University of tunis High Institute
Special Topics in Computer Science
Special Topics in Computer Science NLP in a Nutshell CS492B Spring Semester 2009 Jong C. Park Computer Science Department Korea Advanced Institute of Science and Technology INTRODUCTION Jong C. Park, CS
Customizing an English-Korean Machine Translation System for Patent Translation *
Customizing an English-Korean Machine Translation System for Patent Translation * Sung-Kwon Choi, Young-Gil Kim Natural Language Processing Team, Electronics and Telecommunications Research Institute,
1. Introduction to Spoken Dialogue Systems
SoSe 2006 Projekt Sprachdialogsysteme 1. Introduction to Spoken Dialogue Systems Walther v. Hahn, Cristina Vertan {vhahn,vertan}@informatik.uni-hamburg.de Content What are Spoken dialogue systems? Types
C E D A T 8 5. Innovating services and technologies for speech content management
C E D A T 8 5 Innovating services and technologies for speech content management Company profile 25 years experience in the market of transcription/reporting services; Cedat 85 Group: Cedat 85 srl Subtitle
Bridging CAQDAS with text mining: Text analyst s toolbox for Big Data: Science in the Media Project
Bridging CAQDAS with text mining: Text analyst s toolbox for Big Data: Science in the Media Project Ahmet Suerdem Istanbul Bilgi University; LSE Methodology Dept. Science in the media project is funded
SENTIMENT EXTRACTION FROM NATURAL AUDIO STREAMS. Lakshmish Kaushik, Abhijeet Sangwan, John H. L. Hansen
SENTIMENT EXTRACTION FROM NATURAL AUDIO STREAMS Lakshmish Kaushik, Abhijeet Sangwan, John H. L. Hansen Center for Robust Speech Systems (CRSS), Eric Jonsson School of Engineering, The University of Texas
Annotation in Language Documentation
Annotation in Language Documentation Univ. Hamburg Workshop Annotation SEBASTIAN DRUDE 2015-10-29 Topics 1. Language Documentation 2. Data and Annotation (theory) 3. Types and interdependencies of Annotations
Transcription System Using Automatic Speech Recognition for the Japanese Parliament (Diet)
Proceedings of the Twenty-Fourth Innovative Appications of Artificial Intelligence Conference Transcription System Using Automatic Speech Recognition for the Japanese Parliament (Diet) Tatsuya Kawahara
POSBIOTM-NER: A Machine Learning Approach for. Bio-Named Entity Recognition
POSBIOTM-NER: A Machine Learning Approach for Bio-Named Entity Recognition Yu Song, Eunji Yi, Eunju Kim, Gary Geunbae Lee, Department of CSE, POSTECH, Pohang, Korea 790-784 Soo-Jun Park Bioinformatics
Survey Results: Requirements and Use Cases for Linguistic Linked Data
Survey Results: Requirements and Use Cases for Linguistic Linked Data 1 Introduction This survey was conducted by the FP7 Project LIDER (http://www.lider-project.eu/) as input into the W3C Community Group
Comparative Error Analysis of Dialog State Tracking
Comparative Error Analysis of Dialog State Tracking Ronnie W. Smith Department of Computer Science East Carolina University Greenville, North Carolina, 27834 [email protected] Abstract A primary motivation
Understanding Web personalization with Web Usage Mining and its Application: Recommender System
Understanding Web personalization with Web Usage Mining and its Application: Recommender System Manoj Swami 1, Prof. Manasi Kulkarni 2 1 M.Tech (Computer-NIMS), VJTI, Mumbai. 2 Department of Computer Technology,
CAPTURING THE VALUE OF UNSTRUCTURED DATA: INTRODUCTION TO TEXT MINING
CAPTURING THE VALUE OF UNSTRUCTURED DATA: INTRODUCTION TO TEXT MINING Mary-Elizabeth ( M-E ) Eddlestone Principal Systems Engineer, Analytics SAS Customer Loyalty, SAS Institute, Inc. Is there valuable
Unlocking Value from. Patanjali V, Lead Data Scientist, Tiger Analytics Anand B, Director Analytics Consulting,Tiger Analytics
Unlocking Value from Patanjali V, Lead Data Scientist, Anand B, Director Analytics Consulting, EXECUTIVE SUMMARY Today a lot of unstructured data is being generated in the form of text, images, videos
Shallow Parsing with Apache UIMA
Shallow Parsing with Apache UIMA Graham Wilcock University of Helsinki Finland [email protected] Abstract Apache UIMA (Unstructured Information Management Architecture) is a framework for linguistic
Natural Language to Relational Query by Using Parsing Compiler
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 3, March 2015,
Generating Training Data for Medical Dictations
Generating Training Data for Medical Dictations Sergey Pakhomov University of Minnesota, MN [email protected] Michael Schonwetter Linguistech Consortium, NJ [email protected] Joan Bachenko
Spoken Document Retrieval from Call-Center Conversations
Spoken Document Retrieval from Call-Center Conversations Jonathan Mamou, David Carmel, Ron Hoory IBM Haifa Research Labs Haifa 31905, Israel {mamou,carmel,hoory}@il.ibm.com ABSTRACT We are interested in
CALL-TYPE CLASSIFICATION AND UNSUPERVISED TRAINING FOR THE CALL CENTER DOMAIN
CALL-TYPE CLASSIFICATION AND UNSUPERVISED TRAINING FOR THE CALL CENTER DOMAIN Min Tang, Bryan Pellom, Kadri Hacioglu Center for Spoken Language Research University of Colorado at Boulder Boulder, Colorado
Evaluating grapheme-to-phoneme converters in automatic speech recognition context
Evaluating grapheme-to-phoneme converters in automatic speech recognition context Denis Jouvet, Dominique Fohr, Irina Illina To cite this version: Denis Jouvet, Dominique Fohr, Irina Illina. Evaluating
ONLINE RESUME PARSING SYSTEM USING TEXT ANALYTICS
ONLINE RESUME PARSING SYSTEM USING TEXT ANALYTICS Divyanshu Chandola 1, Aditya Garg 2, Ankit Maurya 3, Amit Kushwaha 4 1 Student, Department of Information Technology, ABES Engineering College, Uttar Pradesh,
Workshop. Neil Barrett PhD, Jens Weber PhD, Vincent Thai MD. Engineering & Health Informa2on Science
Engineering & Health Informa2on Science Engineering NLP Solu/ons for Structured Informa/on from Clinical Text: Extrac'ng Sen'nel Events from Pallia've Care Consult Le8ers Canada-China Clean Energy Initiative
EASY CONTEXTUAL INTENT PREDICTION AND SLOT DETECTION
EASY CONTEXTUAL INTENT PREDICTION AND SLOT DETECTION A. Bhargava, A. Celikyilmaz 2, D. Hakkani-Tür 3, and R. Sarikaya 2 University of Toronto, 2 Microsoft, 3 Microsoft Research ABSTRACT Spoken language
CS 6740 / INFO 6300. Ad-hoc IR. Graduate-level introduction to technologies for the computational treatment of information in humanlanguage
CS 6740 / INFO 6300 Advanced d Language Technologies Graduate-level introduction to technologies for the computational treatment of information in humanlanguage form, covering natural-language processing
The XMU Phrase-Based Statistical Machine Translation System for IWSLT 2006
The XMU Phrase-Based Statistical Machine Translation System for IWSLT 2006 Yidong Chen, Xiaodong Shi Institute of Artificial Intelligence Xiamen University P. R. China November 28, 2006 - Kyoto 13:46 1
2009 Springer. This document is published in: Adaptive Multimedia Retrieval. LNCS 6535 (2009) pp. 12 23 DOI: 10.1007/978-3-642-18449-9_2
Institutional Repository This document is published in: Adaptive Multimedia Retrieval. LNCS 6535 (2009) pp. 12 23 DOI: 10.1007/978-3-642-18449-9_2 2009 Springer Some Experiments in Evaluating ASR Systems
NATURAL LANGUAGE QUERY PROCESSING USING PROBABILISTIC CONTEXT FREE GRAMMAR
NATURAL LANGUAGE QUERY PROCESSING USING PROBABILISTIC CONTEXT FREE GRAMMAR Arati K. Deshpande 1 and Prakash. R. Devale 2 1 Student and 2 Professor & Head, Department of Information Technology, Bharati
GrammAds: Keyword and Ad Creative Generator for Online Advertising Campaigns
GrammAds: Keyword and Ad Creative Generator for Online Advertising Campaigns Stamatina Thomaidou 1,2, Konstantinos Leymonis 1,2, Michalis Vazirgiannis 1,2,3 Presented by: Fragkiskos Malliaros 2 1 : Athens
INF5820, Obligatory Assignment 3: Development of a Spoken Dialogue System
INF5820, Obligatory Assignment 3: Development of a Spoken Dialogue System Pierre Lison October 29, 2014 In this project, you will develop a full, end-to-end spoken dialogue system for an application domain
Comparing Support Vector Machines, Recurrent Networks and Finite State Transducers for Classifying Spoken Utterances
Comparing Support Vector Machines, Recurrent Networks and Finite State Transducers for Classifying Spoken Utterances Sheila Garfield and Stefan Wermter University of Sunderland, School of Computing and
Thai Language Self Assessment
The following are can do statements in four skills: Listening, Speaking, Reading and Writing. Put a in front of each description that applies to your current Thai proficiency (.i.e. what you can do with
Text-To-Speech Technologies for Mobile Telephony Services
Text-To-Speech Technologies for Mobile Telephony Services Paulseph-John Farrugia Department of Computer Science and AI, University of Malta Abstract. Text-To-Speech (TTS) systems aim to transform arbitrary
Sentiment analysis on tweets in a financial domain
Sentiment analysis on tweets in a financial domain Jasmina Smailović 1,2, Miha Grčar 1, Martin Žnidaršič 1 1 Dept of Knowledge Technologies, Jožef Stefan Institute, Ljubljana, Slovenia 2 Jožef Stefan International
Processing: current projects and research at the IXA Group
Natural Language Processing: current projects and research at the IXA Group IXA Research Group on NLP University of the Basque Country Xabier Artola Zubillaga Motivation A language that seeks to survive
Search and Data Mining: Techniques. Text Mining Anya Yarygina Boris Novikov
Search and Data Mining: Techniques Text Mining Anya Yarygina Boris Novikov Introduction Generally used to denote any system that analyzes large quantities of natural language text and detects lexical or
Open Domain Information Extraction. Günter Neumann, DFKI, 2012
Open Domain Information Extraction Günter Neumann, DFKI, 2012 Improving TextRunner Wu and Weld (2010) Open Information Extraction using Wikipedia, ACL 2010 Fader et al. (2011) Identifying Relations for
Machine Learning for natural language processing
Machine Learning for natural language processing Introduction Laura Kallmeyer Heinrich-Heine-Universität Düsseldorf Summer 2016 1 / 13 Introduction Goal of machine learning: Automatically learn how to
Computer-Based Text- and Data Analysis Technologies and Applications. Mark Cieliebak 9.6.2015
Computer-Based Text- and Data Analysis Technologies and Applications Mark Cieliebak 9.6.2015 Data Scientist analyze Data Library use 2 About Me Mark Cieliebak + Software Engineer & Data Scientist + PhD
2014/02/13 Sphinx Lunch
2014/02/13 Sphinx Lunch Best Student Paper Award @ 2013 IEEE Workshop on Automatic Speech Recognition and Understanding Dec. 9-12, 2013 Unsupervised Induction and Filling of Semantic Slot for Spoken Dialogue
Introduction to Pattern Recognition
Introduction to Pattern Recognition Selim Aksoy Department of Computer Engineering Bilkent University [email protected] CS 551, Spring 2009 CS 551, Spring 2009 c 2009, Selim Aksoy (Bilkent University)
USER MODELLING IN ADAPTIVE DIALOGUE MANAGEMENT
USER MODELLING IN ADAPTIVE DIALOGUE MANAGEMENT Gert Veldhuijzen van Zanten IPO, Center for Research on User-System Interaction, P.O. Box 213, 5600 MB Eindhoven, the Netherlands [email protected]
GRAPHICAL USER INTERFACE, ACCESS, SEARCH AND REPORTING
MEDIA MONITORING AND ANALYSIS GRAPHICAL USER INTERFACE, ACCESS, SEARCH AND REPORTING Searchers Reporting Delivery (Player Selection) DATA PROCESSING AND CONTENT REPOSITORY ADMINISTRATION AND MANAGEMENT
AUDIMUS.media: A Broadcast News Speech Recognition System for the European Portuguese Language
AUDIMUS.media: A Broadcast News Speech Recognition System for the European Portuguese Language Hugo Meinedo, Diamantino Caseiro, João Neto, and Isabel Trancoso L 2 F Spoken Language Systems Lab INESC-ID
Micro blogs Oriented Word Segmentation System
Micro blogs Oriented Word Segmentation System Yijia Liu, Meishan Zhang, Wanxiang Che, Ting Liu, Yihe Deng Research Center for Social Computing and Information Retrieval Harbin Institute of Technology,
Effective Self-Training for Parsing
Effective Self-Training for Parsing David McClosky [email protected] Brown Laboratory for Linguistic Information Processing (BLLIP) Joint work with Eugene Charniak and Mark Johnson David McClosky - [email protected]
31 Case Studies: Java Natural Language Tools Available on the Web
31 Case Studies: Java Natural Language Tools Available on the Web Chapter Objectives Chapter Contents This chapter provides a number of sources for open source and free atural language understanding software
CINTIL-PropBank. CINTIL-PropBank Sub-corpus id Sentences Tokens Domain Sentences for regression atsts 779 5,654 Test
CINTIL-PropBank I. Basic Information 1.1. Corpus information The CINTIL-PropBank (Branco et al., 2012) is a set of sentences annotated with their constituency structure and semantic role tags, composed
Open-Source, Cross-Platform Java Tools Working Together on a Dialogue System
Open-Source, Cross-Platform Java Tools Working Together on a Dialogue System Oana NICOLAE Faculty of Mathematics and Computer Science, Department of Computer Science, University of Craiova, Romania [email protected]
Social Media Implementations
SEM Experience Analytics Social Media Implementations SEM Experience Analytics delivers real sentiment, meaning and trends within social media for many of the world s leading consumer brand companies.
Identifying Focus, Techniques and Domain of Scientific Papers
Identifying Focus, Techniques and Domain of Scientific Papers Sonal Gupta Department of Computer Science Stanford University Stanford, CA 94305 [email protected] Christopher D. Manning Department of
TEXT ANALYTICS INTEGRATION
TEXT ANALYTICS INTEGRATION A TELECOMMUNICATIONS BEST PRACTICES CASE STUDY VISION COMMON ANALYTICAL ENVIRONMENT Structured Unstructured Analytical Mining Text Discovery Text Categorization Text Sentiment
