Towards Automatic Animated Storyboarding
|
|
|
- Reynard Arron Foster
- 9 years ago
- Views:
Transcription
1 Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence (2008) Towards Automatic Animated Storyboarding Patrick Ye and Timothy Baldwin Computer Science and Software Engineering NICTA Victoria Laboratories University of Melbourne, Australia University of Melbourne, Australia Abstract In this paper, we propose a machine learning-based NLP system for automatically creating animated storyboards using the action descriptions of movie scripts. We focus particularly on the importance of verb semantics when generating graphics commands, and find that semantic role labelling boosts performance and is relatively robust to the effects of unseen verbs. Introduction Animated storyboards are computer animations created based on movie scripts, and used as crude previews by directors and actors during the pre-production of movies. Creating non-trivial animated storyboards is a time- and labourintensive process, and requires a level of technical expertise that most people do not have. In this research, we propose to automate the process of animated storyboarding using a variety of NLP technologies, potentially saving time and money and also providing a dynamic, visual environment for script creation and fine-tuning. The creation of an animated storyboard can be described as a two-step process. The first step is to construct a static virtual stage with virtual actors and props to approximate the scene to be shot. The second step is to create the interactions between the virtual actors and props, to visualize the events depicted by the action descriptions of the movie scripts. This research is focused on the second step as it is more labourintensive and technically challenging than the first step. There are three major differences between existing NLPaided animation systems and our system. Firstly, most existing systems use handcrafted rules to map the results of language analysis onto graphics commands, whereas our system uses a machine learning system to perform this task automatically. Secondly, existing systems were designed for domain-specific tasks with a controlled vocabulary and syntax, whereas our system is open-domain with no restrictions on the language used other than that the input text is in the style of 3rd person narration. Thirdly, existing systems are all coupled with customised graphics systems designed specifically for their respective tasks, whereas our system is designed to interface with any graphics system that offers access through a programming language style interface. 1 Copyright c 2008, Association for the Advancement of Artificial Intelligence ( All rights reserved. 1 We used a free animation system called Blender3D (www. blender.org) in our experiments. Since the main purpose of animated storyboards is to visualise the actions described in the scripts, and the actions are mostly expressed in verb phrases, the linguistic analysis in our system is focused on verb semantics. Several NLP techniques are used to perform this analysis, namely POS tagging is used to identify the verbs, and semantic role labelling (SRL) (Gildea and Jurafsky 2002) is used to identify the semantic roles of each verb. Furthermore, our linguistic analysis also makes use of lexical resources such as Word- Net (Fellbaum 1998). The main contribution of this research is the proposal of a novel multimodal domain of application for NLP, and the integration of NLP techniques with extra-linguistic and multimodal information defined by the virtual stage. We are especially interested in the usefulness of SRL in our task since it is an intuitively useful NLP technique but has rarely found any practical applications. We are also interested in how well our system could handle verbs not seen in the training data, since rule-based systems often cannot handle such verbs. Related Work Most existing literature on applying NLP techniques to computer animation is limited to domain-specific systems where a customised computer graphics module is driven by a simplistic handcrafted NLP module which extracts features from input text. The pioneering system to support natural language commands to a virtual world was SHRDLU (Winograd 1972), which uses handcrafted rules to interpret natural language commands for manipulating simple geometric objects in a customised block world. The K3 project (Funakoshi, Tokunaga, and Tanaka 2006) uses handcrafted rules to interpret Japanese voice commands to control virtual avatars in a customised 3D environment. Carsim (Johansson et al. 2004; 2005) uses a customised semantic role labeller to extract car crash information from Swedish newspaper text and visually recreate the crash using a customised graphics system. While existing systems have been successful in their own respective domains, this general approach is inappropriate for the task of animated storyboarding because the input language and the domain (i.e. all movie genres) are too broad. Instead, we: (a) use a general-purpose graphics system with a general set of graphics commands not tuned to any one domain, and (b) we use automated NLP and machine learning techniques to perform language analysis on the input scripts, and control the graphics system. 578
2 Our work is also similar to machine translation (MT) in the sense that we aim to translate a source language into a target language. The major difference is that for MT, the target language is a human language, and for our work, the target language is a programming language. Programming languages are much more structured than natural languages, meaning that there is zero tolerance for disfluencies in the target language, as they result in the graphics system being unable to render the scene. Also, the lack of any means of underspecification or ambiguity in the graphics commands means that additional analysis must be performed on the source language to fully disambiguate the language. Overview In this section, we describe our movie script corpus, then provide detailed discussion of the virtual stages and how we use them in our experiments, and finally outline the main computational challenges in automatic animated storyboarding. Movie Script Corpus Movie scripts generally contain three types of text: the scene description, the dialogue, and the action description. In this research, the only section of the script used for language analysis is the acting instructions of the action descriptions, as in (1). 2 That is, we only visualise the physical movements of the virtual actors/props. (1) Andy s hand lowers a ceramic piggy bank in front of Mr. Potato Head and shakes out a pile of coins to the floor. Mr. Potato Head kisses the coins. In real movies, dialogues are of course accompanied by body and facial movements. Such movements are almost never specified explicitly in the movie scripts, and require artistic interpretation on the part of the actual actors. They are therefore outside the scope of this research. We animated around 95% of the Toy Story script using 21 base graphics commands. The script was split up according to its original scene structure. For each annotated scene, we first constructed a virtual stage that resembles the corresponding scene in the original movie. We then performed semantic role labelling of that scene s acting instructions, determined the correct order for visualisation, annotated every visualisable verb with a set of graphics commands and finally, annotated the correct grounding of each argument. In total, 1264 instances of 307 verbs were annotated, with an average of 2.84 base graphics commands per verb token. Virtual Stage A novel and important aspect of this research is that we use the physical properties and constraints of the virtual stage to improve the language analysis of the acting instructions. Below, we outline what sort of information the virtual stage provides. The building blocks of our virtual stages are individual 3D graphical models of real world objects. Each 3D model 2 All the movie script examples in this paper are taken from a modified version of the Toy Story script. is hand-assigned a WordNet synset in a database. The virtual stages are assembled through a drag-and-drop interface between the database and the graphics system. Each 3D model in the virtual stage can be annotated with additional WordNet synsets, which is useful for finding entities that have multiple roles in a movie. For example, the character Woody in Toy Story is both a toy, i.e. an artifact designed to be played with, and a person. Furthermore, each 3D model in the virtual stage can be given one or more name, which is useful for finding entities with more than one title. For example, in the script of Toy Story, the character of Mrs. Davis is often referred to as either Mrs. Davis or Andy s mum. WordNet 2.1 synsets were used to annotate the 3D models. Furthermore, since WordNet provides meronyms for its noun synsets, it will be possible to further annotate the components of each 3D model with the corresponding Word- Net synsets, facilitating the incorporation of more real world knowledge into our system. For example, consider the sentence Andy sits in the chair. Since a chair (typically) consists of a seat, legs, and a back, it would be useful if the system could differentiate the individual components of the 3D model and avoid unnatural visualisation of the sentence, such as Andy sitting on the back of the chair instead of the seat. Finally, an important feature of the virtual stages is that the status (position, orientation, etc) of any virtual object can be queried at any time. This feature makes it possible to extract extra-linguistic information about the virtual objects during the animation process. Main Computational Tasks As stated above, this research centres around extracting verb semantic information from the acting instructions and the virtual stage, and then using this information to create an appropriate animation. We identified two main computational tasks for our research: extracting the timing relationships between the verbs in the input script, and constructing graphics commands using the verb semantic and virtual stage information. Since a correct storyboard requires the actions described by the movie scripts to be animated in the correct chronological order, the extraction of the timing relationships between them is the first task of our system. Generally, verbs should be visualised in the order they appear in the text. For instance, in (1), lower should be visualised first, followed by shake, and finally kiss. However, there are cases were the order of appearance of the verbs does not correspond to the order of visualisation, as in (2) where pummel and bark should be visualised at the same time rather than sequentially in the order they appear. (2) Sid pummels the figure with rocks while Scud is barking wildly at it. The second task of constructing graphics commands is considerably more complicated than the first task, and is the main subject of this paper. This task consists of 2 subtasks: selecting the graphics instructions for the target verbs, and constructing the arguments for the cho- 579
3 sen graphics instructions. The graphics instructions used in this research are similar to procedures in procedural programming languages, and have the general form command(arg0, arg1,..., argn). Therefore, the first subtask is to decide the value of command, and the second subtask is to decide on the values of each argument to the chosen command. The linguistic analysis involved in the task of constructing graphics commands is mostly focused on the extraction of verb semantics. The main NLP technique used here is semantic role labelling. However, as semantic role labellers only identify the surface string associated with each verb argument, an additional step is required to ground the surface string associated with each argument to one or more specific virtual stage object. This grounding task is relatively less studied, and is far from a trivial problem. For instance, consider The toys on the floor in (3). This does not simply correspond to all virtual objects that are annotated with the synset toy 1 in the virtual stage, but has the extra constraints of the objects needing to be on the floor and salient to the current area of focus. (3) The toys on the floor all stop and run towards the monitor. Classifier Setup In this section, we provide details of how we apply machine learning to the task of choosing the graphics commands. Since we use a generic graphics system, it is uncommon for a verb to be fully visualised with a single base graphics command, and very often, it is necessary to use a sequence of combinations of base graphics commands to visualise a verb. Overall, the average number of base graphics commands used to visualise a single verb was In the context of the classifier, we will refer to these combinations of base graphics commands as actions. Base graphics commands within a single action are executed from the same time stamp, sequentially at the beginning of each animation cycle. We treat the subtask of assigning sequences of actions to verbs as a Markov process, and the overall task of action generation as a classification problem. The feature vector of the classifier denotes the latest semantic and virtual stage information, and the class labels denote the next action to be performed. For example, suppose the sentence to be visualised is The man grabs the mug on the table, and the virtual world is configured as in Figure 1a. The classifier should select a set of base commands (i.e. an action) which make the virtual man stand up (Figure 1b), move close to the virtual mug (Figure 1c), and extend his hand to the mug (Figure 1d). Finally, the classifier should select an end-of-sequence action signalling the completion of the visualisation. Each of these movements needs to be executed and the virtual stage updated, and this new information should be used to condition the selection/animation of the next movement. In the remainder of this section, we present the features that we used in our experiments. Linguistic Features The linguistic features used in our system can be divided into the following categories: Verb Features: These features include the original form of the target verb, its lemmas and WordNet synsets, and the hypernyms of these synsets. The inclusion of hypernyms is intended to provide the classifiers with the means to generalise when dealing with previously unseen verbs. The verbs are not disambiguated; instead all WordNet senses of the target verb are included. Collocation Features: These are the lemmas of all the open class words (determined according to the POS tagger output) that occur in the same sentence as the target verb. Semantic Role Features: These features include: the types of roles that the target verb takes (ARG0, ARGM-LOC, etc.); the WordNet synsets of the head words of the grounded constituents combined with the semantic roles of the constituents; and a series of syntactic-collocational features of the constituents of the target verb. Note that we have hand-resolved all anaphora in the data (around 35% of the sentences in the training data contained one or more anaphora), and used this oracle-style anaphora resolution for all methods presented in this paper. The syntactic-collocational features of the constituents were designed to capture the details missed by the grounded semantic roles. The purpose of grounding semantic roles is to find the most relevant virtual object with respect to given textual forms of constituents, which is not sufficient if the most relevant virtual object does not exactly correspond to the semantics of the constituent. For example, consider ARG0 and ARGM LOC of the SRLed sentence in (4). The head of ARG0 is hand which corresponds to a part of the body of Andy. However, since grounding is performed only at the level of virtual objects, ARG0 is grounded to the virtual Andy, and the head of the constituent is lost in the grounding process. Similarly for ARGM LOC, the constituent is a PP, and the most relevant virtual object is the virtual Mr. Potato Head. This is hence what this phrase will be grounded to, resulting in the additional locational information of in front of being lost. The syntactic-collocational features reinstate the linguistic content of each constituent to the feature space, providing the classifier with the means to fine-tune the grounding information and more faithfully render the virtual scene. (4) [ ARG0 Andy s hand] [ TARGET lowers] [ ARG1 a ceramic piggy bank] [ ARG LOC in front of Mr. Potato Head] The syntactic-collocational features are collected recursively from the root of the smallest parse subtree covering all the tokens in the corresponding constituent. Only semantic roles which are prepositional phrases (PP), noun phrases (NP), adjective phrases (ADJP) and adverb phrases (ADVP) are used to generate these features. If the semantic role is a PP, it will generate a feature that includes the head preposition and the recursively retrieved features of its argument. If the constituent is an NP, it will generate a feature that includes the WordNet synset of the 580
4 1a 1b 1c 1d Figure 1: Examples of the sequence of virtual stages corresponding to The man grabs the mug on the table head noun and the recursively retrieved features of its possessor and modifiers. If the semantic role is an ADJP or ADVP, it will generate a feature that includes its head, and all the recursively retrieved features of its modifiers. Take the constituent in front of Mr. Potato Head, for example. Mr. Potato Head would generate the feature of (NP, toy 1 ), and front of Mr. Potato Head would generate the feature of (NP, front 2, (PP MOD, PP, of, (ARG, (NP, toy 1 )))). The ASSERT system (Pradhan et al. 2004) was used to perform the semantic role labelling, the Charniak reranking parser (Charniak and Johnson 2005) was used to perform the parsing, and the SVMTool (Gimnez and Mrquez 2004) POS Tagger (version 1.2.2) was used to POS-tag the input text. Virtual Stage Features The virtual stage features are binary features designed to capture the spatial relationships between the grounded constituents. The general spatial relationships are denoted as the results of different proximity and orientation tests between the virtual objects. For virtual objects that contain bones, the spatial relationships also include the results of the proximity tests between their bones and other virtual objects. In addition, the virtual stage features include description of each ongoing base graphics command that each grounded virtual object is involved in. Experiments The main questions we address in our experiments are: (1) What is the effect of unseen verbs on overall performance? (2) Does semantic role labelling positively contribute to classification accuracy? and (3) How reliant is the classification on accurate grounding? In all the experiments, we divided our original datasets into training, development and test sets. The training set contains roughly the same number of examples as the combined development and test sets, and the development and test sets are of roughly the same size. We used a maximum entropy (Ratnaparkhi 1996) classifier to generate models from the training data. 3 The parameters of these models are then adjusted using the development data, and finally tested on the test data. The details of the parameter tuning can be found in (Ye and Baldwin 2006). To explore the first question of the impact of unseen verbs on overall performance, we performed 2 sets of experiments. 3 The learner is downloadable from: homepages.inf.ed. ac.uk/s /maxent_toolkit.html NP Andy puts S1 S VP NP ADVP piggy the bank back Andy puts the piggy bank back -1 TARGET Figure 2: Example of substitute semantic roles The first set of experiments was designed to test how well the classifiers work when the verbs in the test data are included in the training data whenever possible, i.e. when verbs instances are evenly distributed. The second set of experiments was designed to test how well the classifiers work when all the verbs in the test data are randomly chosen to not appear in the training data. To explore the second question of the impact of SRL, we experimented with three different SRL methods: using gold standard semantic roles, using an automatic labeller (AS- SERT), and using simple parse features as a replacement for semantic roles. The gold standard semantic role data was created by handcorrecting the outputs of ASSERT, and the automatically labelled semantic roles were then evaluated using the evaluation script for the CoNLL 2004 SRL shared task (Carreras and Màrquez 2004). Overall, ASSERT had a precision of.744, a recall of.564 and an F-score of.642. For comparison, over the CoNLL 2004 data, ASSERT had a precision of.724, a recall of.668 and an F-score of.695. It can be observed that the recall over the movie script corpus is significantly lower than the data it was trained for. We also observed that many of the missing semantic roles were caused by the POS tagger, which mis-classified several high frequency verbs (e.g. grab, jump and climb) as nouns. When no semantic roles are used, we used a parse treebased method to extract top-level NPs, PPs, ADJPs and AD- VPs from the output of the Charniak parser, and used the relative positions of these phrases with respect to the target verb as substitute semantic roles. Figure 2 shows the substitute semantic roles for the sentence Andy puts the piggy. 581
5 SRL Precision Recall F-score Gold Auto Table 1: Grounding results bank back. The final question of grounding performance was explored via two sources of grounding information: (1) gold standard grounding annotations based on the gold standard semantic roles; and (2) automatic grounding predictions based on either the gold standard semantic roles, the outputs of ASSERT or the parse tree-based phrase extractor. The automatic grounding predictions were obtained using a string matching method. Given the text descriptor of an object in the text, this method first tries to match it against the user-specified names of the virtual objects, and if no virtual objects match, it then matches the string against the WordNet synsets of the virtual objects. For example, if the virtual stage contains a virtual actor named Woody, and the string Woody or Woody s hand is to be matched, the automatic grounding module will correctly return the virtual Woody. On the other hand, if the virtual stage contains multiple virtual toys, and the string toy is to be matched, the automatic grounding module would incorrectly return all the virtual toys. Table 1 shows the results of the automatic grounding method on the gold standard semantic roles and automatically-labelled semantic roles. We evaluate our models in terms of classification accuracy with respect to the gold-standard action annotations for each verb, on a per-action basis. The majority class baseline for our classifiers is the overall most frequent action (i.e. end-of-sequence). Results and Analysis Experiment Set 1: Balanced Distribution In this first set of experiments, we partitioned the training, development and test datasets such that the annotated verb instances were stratified across the three datasets. The stratification ensures the distributions of verbs in the training, development and test sets are as similar as possible. Our Classifier performance is based on a classifier which has been trained over all the verbs in the training set. Table 2 shows the classification accuracies under different combinations of gold-standard and automatic grounding, and goldstandard, automatic and parser-emulated SRL. The baseline of these classifiers is.386. All our classifiers outperformed the majority class baseline. Unsurprisingly, classifiers trained on both the gold-standard SRL and the gold-standard grounding data achieved the highest accuracy, although the relative difference in including gold-standard SRL was considerably less than was the case with the grounding. Classifier Source of Source of SRL Grounding Gold Auto Parser Gold.568 N/A N/A Auto Table 2: Accuracy for experiment set 1 (no unseen verbs) Classifier Source of Source of SRL Grounding Gold Auto Parser Gold.536 N/A N/A Auto Table 3: Accuracy for experiment set 2 (all unseen verbs) These experiments show that SRL can indeed be useful in the domain of automatic animated storyboarding. It is perhaps slightly surprising that the classifier based on automatically grounded gold-standard semantic roles only slightly outperformed the one based on automatic SRL. A closer look at the results showed that the gold-standard grounding annotation only identified 1266 grounded semantic roles, but the automatic grounding method came up with 1748 grounded semantic roles. This overgeneration of grounded semantic roles introduced a massive amount of noise into the data, thereby lowering the performance of the corresponding classifier. On the other hand, it was observed that ASSERT tended to make the same mistakes for the same verbs. Hence, the same SRL mistakes were consistently present in both the training data and the test data, thereby not causing significant negative effects. These experiments show that there is significant room for improvement in the grounding of the semantic roles. The biggest improvement among the SRL based classifiers came when gold-standard grounding annotation was used. This indicates that the string-matching grounding system we are currently using is inadequate, and deeper linguistic analysis is needed. Experiment Set 2: All Unseen Verbs Recall from above that all the hypernyms of all the senses of the target verbs were included in the verb features, in the hope that they could provide some generalisation power for unseen verbs. The purpose of this second set of experiments is to test how well our classifiers can generalise verb semantics. We first randomly divided our datasets into 4 portions, each containing roughly the same number of verbs, with all instances of a given verb occurring in a single portion. We then performed 4-fold cross-validation, using two portions as training, one as development and one as test data. In our results, we report the average classification accuracy. Table 3 shows the classification accuracies under the same combinations of SRL and grounding data as for experiment set 1. The majority class baseline for these classifiers is.393 All the classifiers in this set of experiments performed worse than in the first set of experiments. This is not sur- 582
6 prising given that we do not have direct access to the semantics of a given verb, as it never occurs directly in the training data. It is also encouraging to see that despite the performance drop, all the classifiers still outperformed the baseline. The classifiers based on automatically-grounded semantic roles suffered the greatest drop in performance: the automatic SRL-based classifier performed below the parseremulated SRL based classifier, even. This is not surprising because most of the verb semantic information used by these classifiers is provided by the grounded semantic roles. In the first set of experiments, even though the semantic role labeller didn t perform well, at least the errors were consistent between the training and test sets, and this consistency to some degree offset the poor semantic role labelling. However, in this set of experiments, the errors in automatically obtained semantic roles are no longer consistent in the training and test sets, and only the correctly-labelled semantic roles generalise well to unseen verbs. This is why the classifier based on the gold-standard grounding annotations suffered the least performance drop among all the semantic role-based classifiers, whereas the classifier based on automatic SRL suffered the most. On the other hand, the parser-emulated semantic roles are not greatly affected by the unseen verbs, because these semantic roles only depend on the relative positions of the relevant phrases to the target verb. These relative positions tend to be much more verb-independent than the real semantic roles, and are therefore less affected by the variation of the target verbs. Discussion and Conclusion Three observations can be made from the experiments. Firstly, unseen verbs have a noticeable negative impact on the overall performance of the classifier, especially when the semantic roles are not of high standard. However, unseen verbs did not cause the classifiers to completely fail, which is encouraging as it shows that our method is relatively robust over unseen verbs, unlike rule-based systems which rely on explicit information for each verb. Secondly, semantic role labelling contributes positively to our task. However, it needs to achieve higher performance in order to be consistently useful when unseen verbs are involved. Thirdly, the performance of the classification relies heavily on the grounding of the semantic roles, and the string matching-based grounding method tends to overgenerate groundings, which in turn introduce noise into our data and reduce the effectiveness of the resultant classifier. In conclusion, we have presented a novel machine learning method for automatically generating animated storyboards. Our method uses several NLP techniques and resources, including parsing, SRL, and WordNet. The experiments presented in this paper show that the features we used are effective and can generalise over previously unseen verbs. Furthermore, we believe that the inclusion of a virtual stage (i.e. encoding of real world interaction) provides a novel perspective to the application and evaluation of NLP techniques, as demonstrated by the use of SRL and parsing in our experiments. Our short term future work will be focused on the grounding process of the semantic roles. We will build a more complete grounding module which is capable of resolving complex NPs and PPs with the aid of the virtual stage, and we will investigate how current techniques in word sense disambiguation, anaphora resolution and co-reference resolution can be incorporated with the grounding process to provide a more integrated solution. References Carreras, X., and Màrquez, L Introduction to the CoNLL-2004 shared task: Semantic role labeling. In Proc. of the 8th Conference on Natural Language Learning (CoNLL-2004), Charniak, E., and Johnson, M Coarse-to-fine n- best parsing and maxent discriminative reranking. In Proceedings of the 43rd Annual Meeting of the Association for Computational Linguistics (ACL 05), Fellbaum, C., ed WordNet: An Electronic Lexical Database. Cambridge, USA: MIT Press. Funakoshi, K.; Tokunaga, T.; and Tanaka, H Conversational animated agent system K3. In Proceedings of 2006 International Conference on Intelligent User Interfaces (IUI2006), Demo Session. Gildea, D., and Jurafsky, D Automatic labeling of semantic roles. Computational Linguistics 28(3): Gimnez, J., and Mrquez, L SVMTool: A general POS tagger generator based on support vector machines. In Proceedings of the 4th International Conference on Language Resources and Evaluation, Johansson, R.; Williams, D.; Berglund, A.; and Nugues, P Carsim: A System to Visualize Written Road Accident Reports as Animated 3D Scenes. In ACL2004: Second Workshop on Text Meaning and Interpretation, Johansson, R.; Berglund, A.; Danielsson, M.; and Nugues, P Automatic text-to-scene conversion in the traffic accident domain. In IJCAI-05, Proceedings of the Nineteenth International Joint Conference on Artificial Intelligence, Pradhan, S.; Hacioglu, K.; Krugler, V.; Ward, W.; Martin, J. H.; and Jurafsky, D Support vector learning for semantic argument classification. Machine Learning 60(1 3): Ratnaparkhi, A A maximum entropy model for partof-speech tagging. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, Winograd, T Understanding Natural Language. Academic Press. Ye, P., and Baldwin, T Semantic role labeling of prepositional phrases. ACM Transactions on Asian Language Information Processing (TALIP) 5(3):
Building a Question Classifier for a TREC-Style Question Answering System
Building a Question Classifier for a TREC-Style Question Answering System Richard May & Ari Steinberg Topic: Question Classification We define Question Classification (QC) here to be the task that, given
Parsing Software Requirements with an Ontology-based Semantic Role Labeler
Parsing Software Requirements with an Ontology-based Semantic Role Labeler Michael Roth University of Edinburgh [email protected] Ewan Klein University of Edinburgh [email protected] Abstract Software
Natural Language Database Interface for the Community Based Monitoring System *
Natural Language Database Interface for the Community Based Monitoring System * Krissanne Kaye Garcia, Ma. Angelica Lumain, Jose Antonio Wong, Jhovee Gerard Yap, Charibeth Cheng De La Salle University
Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information
Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information Satoshi Sekine Computer Science Department New York University [email protected] Kapil Dalwani Computer Science Department
Open Domain Information Extraction. Günter Neumann, DFKI, 2012
Open Domain Information Extraction Günter Neumann, DFKI, 2012 Improving TextRunner Wu and Weld (2010) Open Information Extraction using Wikipedia, ACL 2010 Fader et al. (2011) Identifying Relations for
Effective Self-Training for Parsing
Effective Self-Training for Parsing David McClosky [email protected] Brown Laboratory for Linguistic Information Processing (BLLIP) Joint work with Eugene Charniak and Mark Johnson David McClosky - [email protected]
Chapter 8. Final Results on Dutch Senseval-2 Test Data
Chapter 8 Final Results on Dutch Senseval-2 Test Data The general idea of testing is to assess how well a given model works and that can only be done properly on data that has not been seen before. Supervised
Parsing Technology and its role in Legacy Modernization. A Metaware White Paper
Parsing Technology and its role in Legacy Modernization A Metaware White Paper 1 INTRODUCTION In the two last decades there has been an explosion of interest in software tools that can automate key tasks
CINTIL-PropBank. CINTIL-PropBank Sub-corpus id Sentences Tokens Domain Sentences for regression atsts 779 5,654 Test
CINTIL-PropBank I. Basic Information 1.1. Corpus information The CINTIL-PropBank (Branco et al., 2012) is a set of sentences annotated with their constituency structure and semantic role tags, composed
Identifying Focus, Techniques and Domain of Scientific Papers
Identifying Focus, Techniques and Domain of Scientific Papers Sonal Gupta Department of Computer Science Stanford University Stanford, CA 94305 [email protected] Christopher D. Manning Department of
Comparing Support Vector Machines, Recurrent Networks and Finite State Transducers for Classifying Spoken Utterances
Comparing Support Vector Machines, Recurrent Networks and Finite State Transducers for Classifying Spoken Utterances Sheila Garfield and Stefan Wermter University of Sunderland, School of Computing and
31 Case Studies: Java Natural Language Tools Available on the Web
31 Case Studies: Java Natural Language Tools Available on the Web Chapter Objectives Chapter Contents This chapter provides a number of sources for open source and free atural language understanding software
Author Gender Identification of English Novels
Author Gender Identification of English Novels Joseph Baena and Catherine Chen December 13, 2013 1 Introduction Machine learning algorithms have long been used in studies of authorship, particularly in
What Is This, Anyway: Automatic Hypernym Discovery
What Is This, Anyway: Automatic Hypernym Discovery Alan Ritter and Stephen Soderland and Oren Etzioni Turing Center Department of Computer Science and Engineering University of Washington Box 352350 Seattle,
Interactive Dynamic Information Extraction
Interactive Dynamic Information Extraction Kathrin Eichler, Holmer Hemsen, Markus Löckelt, Günter Neumann, and Norbert Reithinger Deutsches Forschungszentrum für Künstliche Intelligenz - DFKI, 66123 Saarbrücken
Automatic Detection and Correction of Errors in Dependency Treebanks
Automatic Detection and Correction of Errors in Dependency Treebanks Alexander Volokh DFKI Stuhlsatzenhausweg 3 66123 Saarbrücken, Germany [email protected] Günter Neumann DFKI Stuhlsatzenhausweg
Shallow Parsing with Apache UIMA
Shallow Parsing with Apache UIMA Graham Wilcock University of Helsinki Finland [email protected] Abstract Apache UIMA (Unstructured Information Management Architecture) is a framework for linguistic
Testing Data-Driven Learning Algorithms for PoS Tagging of Icelandic
Testing Data-Driven Learning Algorithms for PoS Tagging of Icelandic by Sigrún Helgadóttir Abstract This paper gives the results of an experiment concerned with training three different taggers on tagged
Web 3.0 image search: a World First
Web 3.0 image search: a World First The digital age has provided a virtually free worldwide digital distribution infrastructure through the internet. Many areas of commerce, government and academia have
Domain Independent Knowledge Base Population From Structured and Unstructured Data Sources
Proceedings of the Twenty-Fourth International Florida Artificial Intelligence Research Society Conference Domain Independent Knowledge Base Population From Structured and Unstructured Data Sources Michelle
Automated Extraction of Security Policies from Natural-Language Software Documents
Automated Extraction of Security Policies from Natural-Language Software Documents Xusheng Xiao 1 Amit Paradkar 2 Suresh Thummalapenta 3 Tao Xie 1 1 Dept. of Computer Science, North Carolina State University,
Natural Language to Relational Query by Using Parsing Compiler
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 3, March 2015,
Sentiment Analysis of Movie Reviews and Twitter Statuses. Introduction
Sentiment Analysis of Movie Reviews and Twitter Statuses Introduction Sentiment analysis is the task of identifying whether the opinion expressed in a text is positive or negative in general, or about
Natural Language Processing
Natural Language Processing 2 Open NLP (http://opennlp.apache.org/) Java library for processing natural language text Based on Machine Learning tools maximum entropy, perceptron Includes pre-built models
Brill s rule-based PoS tagger
Beáta Megyesi Department of Linguistics University of Stockholm Extract from D-level thesis (section 3) Brill s rule-based PoS tagger Beáta Megyesi Eric Brill introduced a PoS tagger in 1992 that was based
Visualizing WordNet Structure
Visualizing WordNet Structure Jaap Kamps Abstract Representations in WordNet are not on the level of individual words or word forms, but on the level of word meanings (lexemes). A word meaning, in turn,
Specialty Answering Service. All rights reserved.
0 Contents 1 Introduction... 2 1.1 Types of Dialog Systems... 2 2 Dialog Systems in Contact Centers... 4 2.1 Automated Call Centers... 4 3 History... 3 4 Designing Interactive Dialogs with Structured Data...
Discourse Markers in English Writing
Discourse Markers in English Writing Li FENG Abstract Many devices, such as reference, substitution, ellipsis, and discourse marker, contribute to a discourse s cohesion and coherence. This paper focuses
Predicting the Stock Market with News Articles
Predicting the Stock Market with News Articles Kari Lee and Ryan Timmons CS224N Final Project Introduction Stock market prediction is an area of extreme importance to an entire industry. Stock price is
ScreenMatch: Providing Context to Software Translators by Displaying Screenshots
ScreenMatch: Providing Context to Software Translators by Displaying Screenshots Geza Kovacs MIT CSAIL 32 Vassar St, Cambridge MA 02139 USA [email protected] Abstract Translators often encounter ambiguous
Paraphrasing controlled English texts
Paraphrasing controlled English texts Kaarel Kaljurand Institute of Computational Linguistics, University of Zurich [email protected] Abstract. We discuss paraphrasing controlled English texts, by defining
Semantic annotation of requirements for automatic UML class diagram generation
www.ijcsi.org 259 Semantic annotation of requirements for automatic UML class diagram generation Soumaya Amdouni 1, Wahiba Ben Abdessalem Karaa 2 and Sondes Bouabid 3 1 University of tunis High Institute
ANALYSIS OF LEXICO-SYNTACTIC PATTERNS FOR ANTONYM PAIR EXTRACTION FROM A TURKISH CORPUS
ANALYSIS OF LEXICO-SYNTACTIC PATTERNS FOR ANTONYM PAIR EXTRACTION FROM A TURKISH CORPUS Gürkan Şahin 1, Banu Diri 1 and Tuğba Yıldız 2 1 Faculty of Electrical-Electronic, Department of Computer Engineering
Customizing an English-Korean Machine Translation System for Patent Translation *
Customizing an English-Korean Machine Translation System for Patent Translation * Sung-Kwon Choi, Young-Gil Kim Natural Language Processing Team, Electronics and Telecommunications Research Institute,
11-792 Software Engineering EMR Project Report
11-792 Software Engineering EMR Project Report Team Members Phani Gadde Anika Gupta Ting-Hao (Kenneth) Huang Chetan Thayur Suyoun Kim Vision Our aim is to build an intelligent system which is capable of
Language Interface for an XML. Constructing a Generic Natural. Database. Rohit Paravastu
Constructing a Generic Natural Language Interface for an XML Database Rohit Paravastu Motivation Ability to communicate with a database in natural language regarded as the ultimate goal for DB query interfaces
Transition-Based Dependency Parsing with Long Distance Collocations
Transition-Based Dependency Parsing with Long Distance Collocations Chenxi Zhu, Xipeng Qiu (B), and Xuanjing Huang Shanghai Key Laboratory of Intelligent Information Processing, School of Computer Science,
Overview of MT techniques. Malek Boualem (FT)
Overview of MT techniques Malek Boualem (FT) This section presents an standard overview of general aspects related to machine translation with a description of different techniques: bilingual, transfer,
Efficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words
, pp.290-295 http://dx.doi.org/10.14257/astl.2015.111.55 Efficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words Irfan
Domain Adaptive Relation Extraction for Big Text Data Analytics. Feiyu Xu
Domain Adaptive Relation Extraction for Big Text Data Analytics Feiyu Xu Outline! Introduction to relation extraction and its applications! Motivation of domain adaptation in big text data analytics! Solutions!
Cohesive writing 1. Conjunction: linking words What is cohesive writing?
Cohesive writing 1. Conjunction: linking words What is cohesive writing? Cohesive writing is writing which holds together well. It is easy to follow because it uses language effectively to guide the reader.
Special Topics in Computer Science
Special Topics in Computer Science NLP in a Nutshell CS492B Spring Semester 2009 Jong C. Park Computer Science Department Korea Advanced Institute of Science and Technology INTRODUCTION Jong C. Park, CS
Semantic analysis of text and speech
Semantic analysis of text and speech SGN-9206 Signal processing graduate seminar II, Fall 2007 Anssi Klapuri Institute of Signal Processing, Tampere University of Technology, Finland Outline What is semantic
Blog Post Extraction Using Title Finding
Blog Post Extraction Using Title Finding Linhai Song 1, 2, Xueqi Cheng 1, Yan Guo 1, Bo Wu 1, 2, Yu Wang 1, 2 1 Institute of Computing Technology, Chinese Academy of Sciences, Beijing 2 Graduate School
Modeling coherence in ESOL learner texts
University of Cambridge Computer Lab Building Educational Applications NAACL 2012 Outline 1 2 3 4 The Task: Automated Text Scoring (ATS) ATS systems Discourse coherence & cohesion The Task: Automated Text
Automatic Pronominal Anaphora Resolution in English Texts
Computational Linguistics and Chinese Language Processing Vol. 9, No.1, February 2004, pp. 21-40 21 The Association for Computational Linguistics and Chinese Language Processing Automatic Pronominal Anaphora
B-bleaching: Agile Overtraining Avoidance in the WiSARD Weightless Neural Classifier
B-bleaching: Agile Overtraining Avoidance in the WiSARD Weightless Neural Classifier Danilo S. Carvalho 1,HugoC.C.Carneiro 1,FelipeM.G.França 1, Priscila M. V. Lima 2 1- Universidade Federal do Rio de
Automatic Pronominal Anaphora Resolution. in English Texts
Automatic Pronominal Anaphora Resolution in English Texts Tyne Liang and Dian-Song Wu Department of Computer and Information Science National Chiao Tung University Hsinchu, Taiwan Email: [email protected];
SVM Based Learning System For Information Extraction
SVM Based Learning System For Information Extraction Yaoyong Li, Kalina Bontcheva, and Hamish Cunningham Department of Computer Science, The University of Sheffield, Sheffield, S1 4DP, UK {yaoyong,kalina,hamish}@dcs.shef.ac.uk
Data Mining Yelp Data - Predicting rating stars from review text
Data Mining Yelp Data - Predicting rating stars from review text Rakesh Chada Stony Brook University [email protected] Chetan Naik Stony Brook University [email protected] ABSTRACT The majority
LANGUAGE! 4 th Edition, Levels A C, correlated to the South Carolina College and Career Readiness Standards, Grades 3 5
Page 1 of 57 Grade 3 Reading Literary Text Principles of Reading (P) Standard 1: Demonstrate understanding of the organization and basic features of print. Standard 2: Demonstrate understanding of spoken
International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 5 ISSN 2229-5518
International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November-2013 5 INTELLIGENT MULTIDIMENSIONAL DATABASE INTERFACE Mona Gharib Mohamed Reda Zahraa E. Mohamed Faculty of Science,
Analysis and Synthesis of Help-desk Responses
Analysis and Synthesis of Help-desk s Yuval Marom and Ingrid Zukerman School of Computer Science and Software Engineering Monash University Clayton, VICTORIA 3800, AUSTRALIA {yuvalm,ingrid}@csse.monash.edu.au
CS 6740 / INFO 6300. Ad-hoc IR. Graduate-level introduction to technologies for the computational treatment of information in humanlanguage
CS 6740 / INFO 6300 Advanced d Language Technologies Graduate-level introduction to technologies for the computational treatment of information in humanlanguage form, covering natural-language processing
Language Modeling. Chapter 1. 1.1 Introduction
Chapter 1 Language Modeling (Course notes for NLP by Michael Collins, Columbia University) 1.1 Introduction In this chapter we will consider the the problem of constructing a language model from a set
Clustering of Polysemic Words
Clustering of Polysemic Words Laurent Cicurel 1, Stephan Bloehdorn 2, and Philipp Cimiano 2 1 isoco S.A., ES-28006 Madrid, Spain [email protected] 2 Institute AIFB, University of Karlsruhe, D-76128 Karlsruhe,
Statistical Machine Translation
Statistical Machine Translation Some of the content of this lecture is taken from previous lectures and presentations given by Philipp Koehn and Andy Way. Dr. Jennifer Foster National Centre for Language
Empirical Machine Translation and its Evaluation
Empirical Machine Translation and its Evaluation EAMT Best Thesis Award 2008 Jesús Giménez (Advisor, Lluís Màrquez) Universitat Politècnica de Catalunya May 28, 2010 Empirical Machine Translation Empirical
Low Cost Correction of OCR Errors Using Learning in a Multi-Engine Environment
2009 10th International Conference on Document Analysis and Recognition Low Cost Correction of OCR Errors Using Learning in a Multi-Engine Environment Ahmad Abdulkader Matthew R. Casey Google Inc. [email protected]
Using Knowledge Extraction and Maintenance Techniques To Enhance Analytical Performance
Using Knowledge Extraction and Maintenance Techniques To Enhance Analytical Performance David Bixler, Dan Moldovan and Abraham Fowler Language Computer Corporation 1701 N. Collins Blvd #2000 Richardson,
Presented to The Federal Big Data Working Group Meetup On 07 June 2014 By Chuck Rehberg, CTO Semantic Insights a Division of Trigent Software
Semantic Research using Natural Language Processing at Scale; A continued look behind the scenes of Semantic Insights Research Assistant and Research Librarian Presented to The Federal Big Data Working
Semi-Supervised Learning for Blog Classification
Proceedings of the Twenty-Third AAAI Conference on Artificial Intelligence (2008) Semi-Supervised Learning for Blog Classification Daisuke Ikeda Department of Computational Intelligence and Systems Science,
Extraction of Legal Definitions from a Japanese Statutory Corpus Toward Construction of a Legal Term Ontology
Extraction of Legal Definitions from a Japanese Statutory Corpus Toward Construction of a Legal Term Ontology Makoto Nakamura, Yasuhiro Ogawa, Katsuhiko Toyama Japan Legal Information Institute, Graduate
Towards SoMEST Combining Social Media Monitoring with Event Extraction and Timeline Analysis
Towards SoMEST Combining Social Media Monitoring with Event Extraction and Timeline Analysis Yue Dai, Ernest Arendarenko, Tuomo Kakkonen, Ding Liao School of Computing University of Eastern Finland {yvedai,
PoS-tagging Italian texts with CORISTagger
PoS-tagging Italian texts with CORISTagger Fabio Tamburini DSLO, University of Bologna, Italy [email protected] Abstract. This paper presents an evolution of CORISTagger [1], an high-performance
Semantic Search in Portals using Ontologies
Semantic Search in Portals using Ontologies Wallace Anacleto Pinheiro Ana Maria de C. Moura Military Institute of Engineering - IME/RJ Department of Computer Engineering - Rio de Janeiro - Brazil [awallace,anamoura]@de9.ime.eb.br
Learning Translation Rules from Bilingual English Filipino Corpus
Proceedings of PACLIC 19, the 19 th Asia-Pacific Conference on Language, Information and Computation. Learning Translation s from Bilingual English Filipino Corpus Michelle Wendy Tan, Raymond Joseph Ang,
Symbiosis of Evolutionary Techniques and Statistical Natural Language Processing
1 Symbiosis of Evolutionary Techniques and Statistical Natural Language Processing Lourdes Araujo Dpto. Sistemas Informáticos y Programación, Univ. Complutense, Madrid 28040, SPAIN (email: [email protected])
Stock Market Prediction Using Data Mining
Stock Market Prediction Using Data Mining 1 Ruchi Desai, 2 Prof.Snehal Gandhi 1 M.E., 2 M.Tech. 1 Computer Department 1 Sarvajanik College of Engineering and Technology, Surat, Gujarat, India Abstract
INF5820 Natural Language Processing - NLP. H2009 Jan Tore Lønning [email protected]
INF5820 Natural Language Processing - NLP H2009 Jan Tore Lønning [email protected] Semantic Role Labeling INF5830 Lecture 13 Nov 4, 2009 Today Some words about semantics Thematic/semantic roles PropBank &
Taxonomy learning factoring the structure of a taxonomy into a semantic classification decision
Taxonomy learning factoring the structure of a taxonomy into a semantic classification decision Viktor PEKAR Bashkir State University Ufa, Russia, 450000 [email protected] Steffen STAAB Institute AIFB,
Overview of the TACITUS Project
Overview of the TACITUS Project Jerry R. Hobbs Artificial Intelligence Center SRI International 1 Aims of the Project The specific aim of the TACITUS project is to develop interpretation processes for
Sense-Tagging Verbs in English and Chinese. Hoa Trang Dang
Sense-Tagging Verbs in English and Chinese Hoa Trang Dang Department of Computer and Information Sciences University of Pennsylvania [email protected] October 30, 2003 Outline English sense-tagging
POSBIOTM-NER: A Machine Learning Approach for. Bio-Named Entity Recognition
POSBIOTM-NER: A Machine Learning Approach for Bio-Named Entity Recognition Yu Song, Eunji Yi, Eunju Kim, Gary Geunbae Lee, Department of CSE, POSTECH, Pohang, Korea 790-784 Soo-Jun Park Bioinformatics
Training Paradigms for Correcting Errors in Grammar and Usage
NAACL 10 Training Paradigms for Correcting Errors in Grammar and Usage Alla Rozovskaya and Dan Roth University of Illinois at Urbana-Champaign Urbana, IL 61801 {rozovska,danr}@illinois.edu Abstract This
Comparing Methods to Identify Defect Reports in a Change Management Database
Comparing Methods to Identify Defect Reports in a Change Management Database Elaine J. Weyuker, Thomas J. Ostrand AT&T Labs - Research 180 Park Avenue Florham Park, NJ 07932 (weyuker,ostrand)@research.att.com
Integrating Collaborative Filtering and Sentiment Analysis: A Rating Inference Approach
Integrating Collaborative Filtering and Sentiment Analysis: A Rating Inference Approach Cane Wing-ki Leung and Stephen Chi-fai Chan and Fu-lai Chung 1 Abstract. We describe a rating inference approach
An Approach to Handle Idioms and Phrasal Verbs in English-Tamil Machine Translation System
An Approach to Handle Idioms and Phrasal Verbs in English-Tamil Machine Translation System Thiruumeni P G, Anand Kumar M Computational Engineering & Networking, Amrita Vishwa Vidyapeetham, Coimbatore,
Projektgruppe. Categorization of text documents via classification
Projektgruppe Steffen Beringer Categorization of text documents via classification 4. Juni 2010 Content Motivation Text categorization Classification in the machine learning Document indexing Construction
KNOWLEDGE-BASED IN MEDICAL DECISION SUPPORT SYSTEM BASED ON SUBJECTIVE INTELLIGENCE
JOURNAL OF MEDICAL INFORMATICS & TECHNOLOGIES Vol. 22/2013, ISSN 1642-6037 medical diagnosis, ontology, subjective intelligence, reasoning, fuzzy rules Hamido FUJITA 1 KNOWLEDGE-BASED IN MEDICAL DECISION
Generating Animation from Natural Language Texts and Framework of Motion Database
2009 International Conference on CyberWorlds Generating Animation from Natural Language Texts and Framework of Motion Database Masaki Oshita Kyushu Institute of Technology 680-4 Kawazu, Iizuka, Fukuoka,
An NLP Curator (or: How I Learned to Stop Worrying and Love NLP Pipelines)
An NLP Curator (or: How I Learned to Stop Worrying and Love NLP Pipelines) James Clarke, Vivek Srikumar, Mark Sammons, Dan Roth Department of Computer Science, University of Illinois, Urbana-Champaign.
ARABIC PERSON NAMES RECOGNITION BY USING A RULE BASED APPROACH
Journal of Computer Science 9 (7): 922-927, 2013 ISSN: 1549-3636 2013 doi:10.3844/jcssp.2013.922.927 Published Online 9 (7) 2013 (http://www.thescipub.com/jcs.toc) ARABIC PERSON NAMES RECOGNITION BY USING
Search and Information Retrieval
Search and Information Retrieval Search on the Web 1 is a daily activity for many people throughout the world Search and communication are most popular uses of the computer Applications involving search
Using Feedback Tags and Sentiment Analysis to Generate Sharable Learning Resources
Using Feedback Tags and Sentiment Analysis to Generate Sharable Learning Resources Investigating Automated Sentiment Analysis of Feedback Tags in a Programming Course Stephen Cummins, Liz Burd, Andrew
Automated Extraction of Vulnerability Information for Home Computer Security
Automated Extraction of Vulnerability Information for Home Computer Security Sachini Weerawardhana, Subhojeet Mukherjee, Indrajit Ray, and Adele Howe Computer Science Department, Colorado State University,
An Approach towards Automation of Requirements Analysis
An Approach towards Automation of Requirements Analysis Vinay S, Shridhar Aithal, Prashanth Desai Abstract-Application of Natural Language processing to requirements gathering to facilitate automation
Tagging with Hidden Markov Models
Tagging with Hidden Markov Models Michael Collins 1 Tagging Problems In many NLP problems, we would like to model pairs of sequences. Part-of-speech (POS) tagging is perhaps the earliest, and most famous,
A Method for Automatic De-identification of Medical Records
A Method for Automatic De-identification of Medical Records Arya Tafvizi MIT CSAIL Cambridge, MA 0239, USA [email protected] Maciej Pacula MIT CSAIL Cambridge, MA 0239, USA [email protected] Abstract
Customer Intentions Analysis of Twitter Based on Semantic Patterns
Customer Intentions Analysis of Twitter Based on Semantic Patterns Mohamed Hamroun [email protected] Mohamed Salah Gouider [email protected] Lamjed Ben Said [email protected] ABSTRACT
Managing Variability in Software Architectures 1 Felix Bachmann*
Managing Variability in Software Architectures Felix Bachmann* Carnegie Bosch Institute Carnegie Mellon University Pittsburgh, Pa 523, USA [email protected] Len Bass Software Engineering Institute Carnegie
