LexOnt: A Semi-automatic Ontology Creation Tool for Programmable Web
|
|
|
- Horatio Nash
- 9 years ago
- Views:
Transcription
1 LexOnt: A Semi-automatic Ontology Creation Tool for Programmable Web Knarig Arabshian Bell Labs, Alcatel-Lucent Murray Hill, NJ Peter Danielsen Bell Labs, Alcatel-Lucent Naperville, IL Sadia Afroz Drexel University Philadelphia, PA Abstract We propose LexOnt, a semi-automatic ontology creation tool for a high-level service classification ontology. LexOnt uses the Programmable Web directory as the corpus, although it can evolve to use other corpora as well. The main contribution of LexOnt is its novel algorithm which generates and ranks frequent terms and significant phrases within a PW category by comparing them to external domain knowledge such as Wikipedia, Wordnet and the current state of the ontology. First it matches terms to the Wikipedia page description of the category and ranks them higher, since these indicate domain descriptive words. Synonymous words from Wordnet are then matched and ranked. In a semi-automated process, the user chooses the terms it wants to add to the ontology and indicates the properties to assign these values to and the ontology is automatically generated. In the next iteration, terms within the current state of the ontology are compared to terms in the other categories and automatic property assignments are made for these API instances as well. Introduction In recent years, the availability of services on the World Wide Web has surged. In order to make good use of these services, both human and software consumers require knowledge about existing services. Thus, the need for automatic service discovery and composition is critical. A current technological trend is creating web service mashups. The concept of a mashup is to allow users to create their own content from different types of sources such as websites, RSS Feeds 1, or Flickr 2. A user is able to filter tailored information on a personal page to view and share with others. It is not necessary for a user to know how to create websites, but can do so simply by bringing different components together via a simplified user interface. For example, the Google Maps API 3 is often used in conjunction with location-based web services. Copyright c 2012, Association for the Advancement of Artificial Intelligence ( All rights reserved. 1 RSS Feeds: (accessed Jan 24, 2012) 2 Flickr: (accessed Jan 24, 2012) 3 Google Maps API: (accessed Jan 24, 2012) Currently, there are a number of web services in various domains, such as social media or mapping services that offer their APIs to be used in mashup applications. Programmable Web 4 is one such directory that offers a listing of Web service APIs. Programmable Web classifies APIs in a flat categorization where each API is manually classified within a single service category. Search is limited to attributes such as protocol or messaging type and is not related to semantic attributes of the service category. Thus, service discovery and composition within the ProgrammableWeb directory is a challenging process, since it requires considerable manual effort to locate services, understand their capabilities and compose mashup applications. Furthermore, every site has its databases modeled in a specific way, causing semantically equivalent properties to be defined differently, since data is not easily shared across different domains in the Internet. We enhance the service descriptions by using an ontology to describe each service category. The main benefits in using an ontology to describe services are that it provides structure and semantics to service metadata which allows it to be easily classified and shared. With automated classification, service discovery becomes seamless even as more services are added or the ontology changes. Also, services can be discovered according to specific attributes, which is useful when composing services in mashup applications. Furthermore, service metadata can be shared across different domains. If a service description for a Travel or Mapping service is stored in a generic ontology repository, a mashup application can easily compose these types of services by finding its description within a repository. Thus, with an ontology description, an API can be automatically classified and queried by its attributes resulting in automated service discovery and composition. One of the difficulties in using ontologies for describing a service domain is in creating a generic classification that describes the service class and its high-level properties. Currently, most of the semi-automated ontology generation tools either have an established taxonomy or a structured corpus, such as in the fields of biology and medicine which aid domain experts in creating ontologies. Furthermore, the 4 Programmable Web: (accessed Jan 24, 2012)
2 tools that analyze generic text to semi-automatically create an ontology, create taxonomic hierarchies but do not indicate property descriptions of the classes. We discuss this in the related works section. As we have indicated above, there is a need for high-level classification ontologies of different service classes or domains and their high-level properties but the current state of the art does not accomplish this. Thus, in order to make the use of ontologies for web service descriptions prevalent, a tool is required that will allow anyone to create a high-level ontology description of a domain given a corpus of data. We propose LexOnt, a semi-automatic ontology creation tool that suggests high-level property terms for a given service class which distinguish it from the rest of the categories. Although we have tested this tool using the PW service directory, we anticipate that LexOnt can be applied to all types of unstructured text corpora describing a certain domain. The main contribution of LexOnt is its novel algorithm which takes as its input unstructured text of a given domain and generates and ranks frequent terms and significant phrases within a PW category as candidate object property assignments by comparing them to external domain knowledge within Wikipedia 5, Wordnet 6 and the current state of the ontology. LexOnt can thus be used to semi-automatically create ontologies even if the ontology engineer is not necessarily an expert of a certain domain, since the terms and phrases it suggests to the user are those that are common within the corpus and the domain. LexOnt accomplishes this by initially generating a list of terms and phrases obtained by well-known NLP algorithms such as TF-IDF (Salton and Buckley 1988) and significant phrase generation. It then matches these terms to the Wikipedia page description of the category and Wordnet synonyms and ranks them higher, since these indicate domain descriptive words. In a semi-automated process, the user chooses the terms it wants to add to the ontology and indicates the properties to assign these values to and LexOnt generates the appropriate ontology entities. In the next iteration, terms within the current state of the ontology are compared to terms in the other categories and automatic property assignments are made for these API instances as well. We describe the details of LexOnt s algorithm, implementation and results in this paper. Related Work Most of the related work closest to our work involves semiautomated ontology creation for taxonomic hierarchies or domains that already have some kind of structural description. Machine learning and NLP techniques are used on text corpora alongside an already existing ontological description of the domain to generate an ontology for that specific dataset. There is work that uses clustering for semi-automatic construction of ontologies from parsed text corpora (Bisson, Nedellec, and Canamero 2000), (Reinberger and Spyns 2004), for creating taxonomic hierarchies (P. Cimiano 2004) and for generating topic ontolo- 5 Wikipedia: (accessed Jan 24, 2012) 6 Wordnet: (accessed Jan 24, 2012) gies (B. Fortuna 2007). The work closest to ours are those that involve finding property relationships between concepts or use external knowledge bases. A few systems which do either have been proposed: TextToOnto (Maedche and S. Staab 2000), OntoLT (Buitelaar and Olejnik 2004), and Ontolearn (Navigli and P. Velardi 2003). Text2Onto creates an ontology from annotated texts. This system incorporates probabilistic ontology models (POMs). It shows a user different models ranked according to the certainty ranking and does linguistic preprocessing of the data. It also finds properties that distinguish a class from another. The main difference with LexOnt is that Text2Onto uses an annotated corpus for term generation. OntoLT allows a user to define mapping rules which provides a precondition language that annotates the corpus. Preconditions are implemented using XPATH expressions and consist of terms and functions. According to the preconditions that are satisfied, candidate classes and properties are generated. Again, OntoLT uses pre-defined rules to find these relationships. OntoLearn, like LexOnt, uses an unstructured corpus and external knowledge of natural language definitions and synonyms to generate terms. However, the ontology that is generated is a hierarchical classification and does not involve property assertions. Thus, the main difference between LexOnt and these systems is that they either start off with some kind of annotated description in addition to the text corpora or they produce a pure hierarchical ontological description and exclude high-level property descriptions within the service domain. LexOnt, on the other hand, starts with unstructured text of a given service domain and uses external knowledge from online sources such as Wikipedia and Wordnet to generate terms. It then uses the terms from the constructed ontology to influence the generation of future terms. Also, in our work, we are concentrating mainly on describing service domains. Although our techniques may be applied to all types of domains, it works best for describing high-level service domains where generic properties are shared across the single domain. Programmable Web Overview Currently ProgrammableWeb maintains a collection of over 5000 APIs divided into 56 high-level categories. The API categories are service classes such as: Advertising, Answers, Blog Search, Blogging, Bookmarks, Calendar, Chat, Database, Dating, Dictionary, , Photo, Real Estate, Social, Travel, Utility, Video, Weather, etc. A PW user registers its service by submitting a registration form. The form contains, among other things, URLs for the provider and sample links.. After reviewing the API, the PW team manually assigns it to a single category. PW generates two HTML pages, given an API registration: provider.html is the front page of the provider s URL and sampleurl.html is an HTML page published in the PW website which has specific information for the API itself. Throughout the paper, we denote the textual corpus as the API description. The API descriptions is natural language text without any fixed structure, since we used the text data stripped from the HTML pages of an API.
3 In order to search the directory, a user can either enter tags or search via a form. The form provides a drop-down menu to choose the category and properties such as keywords, data format, company, protocols. These properties, although helpful in somewhat filtering out APIs, do not really help in searching for a service based on its semantic description. Figure 1 gives an overview of the PW search site. The search results are usually a long list of links to possible API matches and a brief description of each API. If the user has a specific service in mind or wants to create a mashup of a few services, he will have to browse through the API site to understand what the API is about before he even begins to read through the documentation to start coding. Thus, merely the discovery process can be quite lengthy. Figure 1: Programmable Web Directory Search LexOnt: A Semi-automatic Ontology Creation Tool for Programmable Web Motivation The goal of LexOnt is to find high-level properties of a class that distinguish it from other classes. Specifically, we are using the LexOnt tool to determine generic properties of service classes in the Programmable Web directory such that they can be categorized and searched for automatically using an ontology classification. The service classification can be likened to the profile section of an OWL-S (OWL for Services) (owl ) ontology where a service s semantics is described in a high-level ontology. For future work, we will look at describing the full service functionality of the PW services. To illustrate, take the Travel and Social service categories. A typical web service that offers travel information may have features to book flights, hotels, or cars. Other travel sites may provide tourism advice or aggregate travel data and compare prices. Social networking services typically have user profiles, friend lists and some kind of sharing attribute. There are many social networking services available on the Internet today and these are distinguished by what is being shared: photos, videos, tags, reviews, etc. Thus, the goal of LexOnt is to enable a user to distinguish common terms within a category such that these terms can be applied as distinguishing features of a service domain. Once these terms are chosen, they are added as hasfeature properties to the ontology. It is unrealistic to assume that an ontology can be created in a purely automated fashion. Ontologies are often subjective descriptions of a given domain which require human evaluation. However, we can semi-automate this process by analyzing a corpus of data within a domain and aiding a user who may not be completely familiar with a domain to choose terms that may describe this domain, and then automating attribute creation. Furthermore, as the ontology is being created for a domain, the information within it can also be used to give higher weights for the terms within the corpus that match the terms that have been instantiated in the ontology. Algorithms The novelty in LexOnt s algorithm is twofold: 1) use information from HTML text describing service APIs, Wikipedia articles describing that service domain and Wordnet for synonymous terms as our corpus for generating a top-n list of words and phrases that may be used toward assigning distinguishing service features 2) semi-automate the construction of the ontology by labeling terms that have been assigned manually to the ontology based on the term generation and then creating or asserting property assignments within the ontology for subsequent iterations. Thus, we incorporate the current state of the ontology into the corpus of data itself to assign higher weights or labels to those terms which are already assigned in the ontology and then regenerate a list of terms from the text data given the current ontological terms. Term Extraction with TF-IDF and Significant Phrases Initially, LexOnt uses well-established NLP techniques to extract terms from the API text descriptions that may be used as distinguishing features of the service. Two lists are generated: a top N list of TF-IDF ranked terms and a list of two-termed significant phrases. TF-IDF: Text frequency- inverse document frequency (TF-IDF) score of a word in a corpus shows how important the word is in the corpus. Importance of a word in a particular document depends on how frequently the word has been used in the document and how common the word is in all the documents in the corpus. Significant Phrases: A signficant phrase consists of two or more words. We chose to generate a list of these phrases, in addition to single term TF-IDF ranked words, because we determined in our tests that some of these phrases gave a very good indication of high-level property descriptions. For example, in the Advertising service class, significant terms generated were things like: mobile advertising, facebook advertising or campaign advertising. This immediately gave a synopsis of what the API was about. Significant
4 phrase generation is a two-phase process. First, collocations, terms that appear together, are determined. Then, from this list, unique collocations are filtered out. The Chi-square test calculates the significance of the collocated words. It measures how often the words in a phrase are seen together and how often they are seen alone. For example, if in a description the word social appears 8 times, stream appears 8 times and social stream appears 8 times, then social stream is considered as a significant phrase as there is a high correlation of these words to be appeared together as a phrase. To calculate chi-square probability of an n length phrase, a n-by-n table is constructed. The χ 2 sums the differences between observed and expected values in all squares of the table. Once the collocations are determined, we filter this list and find the unique phrases to determine the distinct properties. This helps to prune out irrelevant word collocations such as: sign in, api usage, log out that appear in almost all the API descriptions and are not important for our purpose. We perform the following steps to find distinctive phrases of an API: 1. Create testing sets and training sets. The testing set is generated via the API being processed. The training sets, on the other hand, are generated using all the APIs that do not have the same category as the API being processed. For example, if the API being processed belongs to the Advertising category, choose APIs which are not from the Advertising category. 2. Find frequencies of n-grams in the training set. 3. Find frequencies of n-grams in the testing set. 4. N-grams in the testing set are sorted according to their significant score. The significant score is the z score for binomial distribution 7. Use of external knowledge-base: Wikipedia, Wordnet and Constructed Ontology Given the categories of an API, external resources, like Wikipedia, Wordnet and the constructed ontology can be used to leverage the underlying concepts and properties of an API. For example, top 20 words from the Advertising Wikipedia page are: advertising, marketing, brand, television, semiotics, advertisement, billboard, radio, product, bowl, sponsor, consumer, advertise, placement, super, logo, commercial, infomercial, message, promotion. LexOnt uses these words to rank the top terms. For each of these terms, it then looks for synonymous terms that may be related to it. It uses Wordnet to find synonyms of the top-n words that are being used. Finally, it searches through the ontology to see if there are matching terms and if there are, labels or ranks these terms to indicate that they have already been created. Below, we illustrate the step-bystep algorithm for using the external knowledgebase. Table 1 also shows an example of generated terms for one of the APIs in the Advertising category. 1. Extract Wikipedia page for each category. 7 z-score for binomial distribution: (accessed Jan 24, 2012) Table 1: Example of property selection: 140 Proof Top N TF- Advertising, marketing, brand, IDF from television, semiotics, advertisement, Wiki billboard, radio, product, bowl, sponsor, consumer, advertise, placement, super, logo, commercial, infomercial Top N TF- ad, advertisement, advertizement, IDF from advertising, advertizing, advert, Wordnet promotion, direct mail, preview, advertorial, mailer, newspaper ad, commercial, circular, teaser, top billing Top N TF- proof, Persona, Stream, Replies, IDF from Authors, Say, hello, Ad, Brands, social, PW Categorence, Consumers, Advertisers, Audi- Ads Top N TF- Advertisers (wiki), Consumers IDF Ranked (wiki), Social (wiki), Brands based on external kb (wiki), Ads (related), Ad (related), proof, Persona, Stream, Replies, Top N Significant Phrases ranked based on external kb Authors, Say, Hello, Audience stream advertising (wiki), social stream (wiki), say hello, author, replies, google groups, ober, michaels, proof, erik michaels, persona targeting 2. Find top words based on TF-IDF ranking. 3. If a word or phrase in the API contains any of the top Wikipedia words, rank it higher than the others. 4. Find synonymous or related terms to the list of generated terms using Wordnet 5. If a word or phrase in the API contains any of the related terms, rank and label these 6. If any of the generated terms lexically match terms in the ontology, label these as well. Semi-automatic ontology creation process LexOnt is implemented as a Protege plugin to provide a user-friendly interface during the interactive process of creating and editing an ontology. Thus, the user installs the LexOnt tab in the Protege editor and can seamlessly switch to the ontology editor tab within the Protege application. The diagram in Figure 2 illustrates the process of ontology creation with LexOnt. It starts with a set of PW service categories, generates a top-n list of terms and phrases, given the initial corpus of data and creates a list of ontology classes based on the PW service categories. The user then creates a few high-level object properties. In our case, we have chosen an object property that indicates distinguishing features of a service. Thus, we have created a F eature class and its corresponding hasf eature property, which points to this class. The PW classes and their APIs are displayed in a list on the side bar. When the user clicks on an API, top-n terms
5 and phrases appear. The user can then choose relevant terms and enter it into the system as property-value assignments. We have built the interface in such a way that when the user right clicks on a term, he can either search for it in the original corpus to see how it is used or assign it as a possible property value in the ontology. For example, if the user sees that the phrase mobile advertising is relevant, it can right click on the term, choose the appropriate property to assign this term to, which is hasf eature and then click on the Find button. This displays all the locations in the API which have this term in it. The user can also do a search for this term in all the other APIs within a category. Matches within all the APIs are displayed in a table that can be edited by the user. Once the user chooses the terms that best describe the APIs and edits it, it clicks on the Create button which automatically creates instances of APIs within that category and assigns the relevant terms chosen to the hasf eature property. Figure 2: Ontology Creation Process with LexOnt Example An example of property assignments for the 140P roof API instance is illustrated in the snapshot in Figure 3. The figure shows that the 140Proof API instance belongs to the Advertising Service class and has features from the Social Stream Advertising Feature and Persona Targeting Advertising Feature classes which are both subclasses of the Feature class. This example not only shows how the instances are being assigned, but how the Advertising Feature class has been constructed. As terms are chosen from the corpus and assigned as properties to individuals, the Advertising Feature class is constructed automatically as well. Thus, it can be seen that LexOnt has determined that the Advertising Service class can have advertising features such as banner, campaign, cloud, contextual, facebook, etc. Figure 3: Ontology snapshot for the 140P roof API Instance Implementation and Evaluation LexOnt was implemented using three different Java APIs. The Lingpipe API 8 was used for the NLP algorithms to generate TF-IDF terms and Significant Phrases. The Protege plugin GUI was implemented using the Protege API (Knublauch et al. 2004). Finally, the ontology generation code was implemented using the latest version of the OWL-API (Bechhofer, Volz, and Lord 2003) which also handles the next generation of OWL 2 (Grau et al. 2008). As a future paper, we will discuss the contributions we have made in creating a user-friendly interface for semi-automatic ontology creation. Currently, only OntoLT has been implemented as a Protege plugin, but is now incompatible with the current version of Protege. Since Protege is a robust ontology editor, having LexOnt as a Protege plugin tab has significantly improved the ontology creation process. We tested for three things when evaluating LexOnt: the precision/recall of the TF-IDF term and Significant Phrase generation; how helpful the external knowledge base was when choosing terms; and whether or not the terms were used in their exact form, similar form or different forms. The results are seen in tables 2, 3, 4 respectively. The categories that were used to generate the terms were chosen according to specificity, number of APIs and a priori knowledge of the domain. The Advertising and Real Estate service categories each had a manageable number of APIs, on average around 40 APIs. These categories are also specifically defined domains that had comprehensive Wikipedia pages. Additionally, the co-author who created the ontology for these categories did not have much knowledge about either of these domains. We then chose two categories that the ontology creator was familiar with: Travel and Social. The former had around 75 APIs and the latter had over 200 APIs. Finally, we chose a generic Utility category which had 65 APIs and had a hodgepodge of various services with no corresponding Wikipedia page. 8 Alias-i LingPipe (accessed Jan 24, 2012
6 Table 4: Term Usage Category Equal Terms Similar Terms Different Terms Advertising 85.71% 100% 65.71% Real Estate 16.67% 91.67% 66.67% Social 1.73% 86.9% 79.1% Travel 6.25% 100% 2.3% Utility 5% 60% 50% Table 2: Precision/Recall Stats Category Sig. TF- Recall Phrase IDF Advertising 3.98% 2.77% 43.88% Real Estate 1.02%.92% 9.57% Social 3.21% 2.8% 20.19% Travel 1.96% 2.4% 30.91% Utility 9.58% 3.83% 34.91% Table 3: Percentage of terms used from KB Category Sig. Phrase TF- IDF Advertising 41.38% 52.73% Real Estate 100% 100% Social 31.90% 11.38% Travel 82.26% 72.73% Utility 0% 0% Table 2 showed that on average, the precision of generated terms was around 4% for Significant Phrases and 3% for TF-IDF terms. The recall for both was around 28%. Although these numbers are not very promising in and of themselves, when looking at Tabel 3 to see how these match to the external knowledge base, the results are quite promising. For the categories with external KBs, the percentage of terms matching KB terms were between 30 and 100 percent for Significant Phrases and 11 and 100 percent for TF-IDF terms. We also determined that for the categories with a smaller API set and well-defined Wikipedia pages, these numbers were quite high. Thus, it was quite feasible to quickly assess distinguishing features of a category just by looking at the generated terms and how well they matched the external KB. For the Social category, the percentage of terms used from a KB was not as high because there was a lot of variance within this cateogry. Social APIs varied widely between what was shared. Aside from the usual friends, blogs, statuses, tags, photos and videos, there were APIs that shared other types of items such as wine lists, books, courses or fashion. The Utility category didn t have a corresponding Wikipedia page and thus we relied heavily on the TF-IDF and Significant phrase generation and finding these terms within the original corpus via the LexOnt interface in order to determine what the API features were. Finally, we tested to see how these terms were actually assigned within the instances. Thus, we checked to see whether these matches were exact, similar or completely different. For example, if LexOnt produced a term mobile but the actual ontology assignment was mobile advertising, this would count as a similar match. From our results, we saw that the percentage of equal and similar matches for individuals were quite high for all the categories. However, the percentage of different matches varied. Again, we can see that for a category that has a wide variety of data, such as in Social, the percentage of differing terms ranked higher than the others. Discussion LexOnt has shown to be an effective tool for semi-automated ontology creation. From our initial results, we have determined that using an external knowledge base to filter out generated terms and phrases greatly increases the accuracy of the feature selection. It also helps in understanding the common terms within a corpus. We will continue to test Lex- Ont for the other categories and compare results. Future Work and Conclusion For future work, we would like to test LexOnt on other types of corpora besides the PW directory to see how well it works within other domains such as in medicine or biology. We also want improve the seamless integration of LexOnt s automated ontology creation mechanism with the user s interaction. For example, when a user is in the process of creating an ontology, he should be able to save the current state of the LexOnt tool and the tool should be aware of all the instances created, and store information not only about generated terms that were accepted but those that were rejected as well. This way, LexOnt will avoid false positives within the generated terms. Additionally, once we have a complete ontology, we will build a querying interface to Programmable Web and test to see how service discovery has improved with the new ontology. Eventually, we would like to describe services not just on a high-level but in its functional details. Allowing standard descriptions of a service s inputs, outputs, pre and post conditions will aid in autmomatic service composition. Once we have the high-level discovery phase automated with this ontology, we plan on continuing this work to see how we can generate an ontology description of the service which include its functional details. In conclusion, we have presented LexOnt, a semiautomatmic ontology generator that aids in the ontology creation of a high-level service ontology. It uses the Programmable Web directory of services, Wikipedia, Wordnet and the current state of the generated ontology to suggest relevant terms that may be incorporated within the ontology. LexOnt builds the ontology iteratively, by interacting with the user, taking in terms that the user has chosen, adding these to the ontology and ranking terms according to the external knowledge base. From our initial findings, we have determined that LexOnt is a useful tool to generate a highlevel ontology description of a domain, specifically for users who are not domain experts.
7 References B. Fortuna, M. Grobelnik, D. M Ontogen: Semiautomatic ontology editor. In Human Interface, Part II, HCII Bechhofer, S.; Volz, R.; and Lord, P Cooking the semantic web with the owl api. The Semantic Web-ISWC Bisson, G.; Nedellec, C.; and Canamero, L Designing clustering methods for ontology building: The mok workbench. In Ontology Learning Workshop, The 14th European Conference on Artificial Inteligence (ECAI). Buitelaar, S., and Olejnik, D A protege plug-in for ontology extraction from text based on linguistic analysis. Grau, B.; Horrocks, I.; Motik, B.; Parsia, B.; Patel- Schneider, P.; and Sattler, U Owl 2: The next step for owl. Web Semantics: Science, Services and Agents on the World Wide Web 6(4): Knublauch, H.; Fergerson, R.; Noy, N.; and Musen, M The protégé owl plugin: An open development environment for semantic web applications. The Semantic Web ISWC Maedche, A., and S. Staab, S Semi-automatic engineering of ontologies from text. In 12th International Conference on Software Engineering and Knowledge Engineering. Navigli, R., and P. Velardi, A. G Ontology learning and its application to automated terminology translation. In IEEE Intelligent Systems, vol. 18:1. Owl-s (semantic markup for web services). P. Cimiano, A. Pivk, L. S.-T. S. S Learningtaxonomicrelations from heterogeneous evidence. In Ontology Learning and Population Workshop The 16th European Conference on Artificial Inteligence (ECAI). Reinberger, M., and Spyns, P Discovering knowledge in texts for the learning of dogma-inspired ontologies. In In Proceedings of the Ontology Learning and Population Workshop, The 16th European Conference on Artificial Inteligence (ECAI). Salton, G., and Buckley, C Term-weighting approaches in automatic text retrieval. Information processing & management 24(5):
Chapter 6. Attracting Buyers with Search, Semantic, and Recommendation Technology
Attracting Buyers with Search, Semantic, and Recommendation Technology Learning Objectives Using Search Technology for Business Success Organic Search and Search Engine Optimization Recommendation Engines
Fogbeam Vision Series - The Modern Intranet
Fogbeam Labs Cut Through The Information Fog http://www.fogbeam.com Fogbeam Vision Series - The Modern Intranet Where It All Started Intranets began to appear as a venue for collaboration and knowledge
Search and Information Retrieval
Search and Information Retrieval Search on the Web 1 is a daily activity for many people throughout the world Search and communication are most popular uses of the computer Applications involving search
Folksonomies versus Automatic Keyword Extraction: An Empirical Study
Folksonomies versus Automatic Keyword Extraction: An Empirical Study Hend S. Al-Khalifa and Hugh C. Davis Learning Technology Research Group, ECS, University of Southampton, Southampton, SO17 1BJ, UK {hsak04r/hcd}@ecs.soton.ac.uk
Incorporating Window-Based Passage-Level Evidence in Document Retrieval
Incorporating -Based Passage-Level Evidence in Document Retrieval Wensi Xi, Richard Xu-Rong, Christopher S.G. Khoo Center for Advanced Information Systems School of Applied Science Nanyang Technological
IT services for analyses of various data samples
IT services for analyses of various data samples Ján Paralič, František Babič, Martin Sarnovský, Peter Butka, Cecília Havrilová, Miroslava Muchová, Michal Puheim, Martin Mikula, Gabriel Tutoky Technical
How To Make Sense Of Data With Altilia
HOW TO MAKE SENSE OF BIG DATA TO BETTER DRIVE BUSINESS PROCESSES, IMPROVE DECISION-MAKING, AND SUCCESSFULLY COMPETE IN TODAY S MARKETS. ALTILIA turns Big Data into Smart Data and enables businesses to
Understanding Web personalization with Web Usage Mining and its Application: Recommender System
Understanding Web personalization with Web Usage Mining and its Application: Recommender System Manoj Swami 1, Prof. Manasi Kulkarni 2 1 M.Tech (Computer-NIMS), VJTI, Mumbai. 2 Department of Computer Technology,
PDF hosted at the Radboud Repository of the Radboud University Nijmegen
PDF hosted at the Radboud Repository of the Radboud University Nijmegen The following full text is an author's version which may differ from the publisher's version. For additional information about this
Ontology construction on a cloud computing platform
Ontology construction on a cloud computing platform Exposé for a Bachelor's thesis in Computer science - Knowledge management in bioinformatics Tobias Heintz 1 Motivation 1.1 Introduction PhenomicDB is
Domain Independent Knowledge Base Population From Structured and Unstructured Data Sources
Proceedings of the Twenty-Fourth International Florida Artificial Intelligence Research Society Conference Domain Independent Knowledge Base Population From Structured and Unstructured Data Sources Michelle
INTERNET MARKETING. SEO Course Syllabus Modules includes: COURSE BROCHURE
AWA offers a wide-ranging yet comprehensive overview into the world of Internet Marketing and Social Networking, examining the most effective methods for utilizing the power of the internet to conduct
A Case Study of Question Answering in Automatic Tourism Service Packaging
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 13, Special Issue Sofia 2013 Print ISSN: 1311-9702; Online ISSN: 1314-4081 DOI: 10.2478/cait-2013-0045 A Case Study of Question
Semantic Search in Portals using Ontologies
Semantic Search in Portals using Ontologies Wallace Anacleto Pinheiro Ana Maria de C. Moura Military Institute of Engineering - IME/RJ Department of Computer Engineering - Rio de Janeiro - Brazil [awallace,anamoura]@de9.ime.eb.br
Term extraction for user profiling: evaluation by the user
Term extraction for user profiling: evaluation by the user Suzan Verberne 1, Maya Sappelli 1,2, Wessel Kraaij 1,2 1 Institute for Computing and Information Sciences, Radboud University Nijmegen 2 TNO,
Search Result Optimization using Annotators
Search Result Optimization using Annotators Vishal A. Kamble 1, Amit B. Chougule 2 1 Department of Computer Science and Engineering, D Y Patil College of engineering, Kolhapur, Maharashtra, India 2 Professor,
Clustering Technique in Data Mining for Text Documents
Clustering Technique in Data Mining for Text Documents Ms.J.Sathya Priya Assistant Professor Dept Of Information Technology. Velammal Engineering College. Chennai. Ms.S.Priyadharshini Assistant Professor
TDPA: Trend Detection and Predictive Analytics
TDPA: Trend Detection and Predictive Analytics M. Sakthi ganesh 1, CH.Pradeep Reddy 2, N.Manikandan 3, DR.P.Venkata krishna 4 1. Assistant Professor, School of Information Technology & Engineering (SITE),
A quick guide to. Social Media
A quick guide to Social Media In this guide... Learn how to integrate your email marketing with social media to get the most out of online buzz! Use Twitter and Facebook integrations to enable readers
Lecture Overview. Web 2.0, Tagging, Multimedia, Folksonomies, Lecture, Important, Must Attend, Web 2.0 Definition. Web 2.
Lecture Overview Web 2.0, Tagging, Multimedia, Folksonomies, Lecture, Important, Must Attend, Martin Halvey Introduction to Web 2.0 Overview of Tagging Systems Overview of tagging Design and attributes
An Approach to support Web Service Classification and Annotation
An Approach to support Web Service Classification and Annotation Marcello Bruno, Gerardo Canfora, Massimiliano Di Penta, and Rita Scognamiglio [email protected], [email protected], [email protected],
Practical Semantic Web Tagging and Tag Clouds 1
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 6, No 3 Sofia 2006 Practical Semantic Web Tagging and Tag Clouds 1 Ivo Marinchev Institute of Information Technologies, 1113
CONCEPTCLASSIFIER FOR SHAREPOINT
CONCEPTCLASSIFIER FOR SHAREPOINT PRODUCT OVERVIEW The only SharePoint 2007 and 2010 solution that delivers automatic conceptual metadata generation, auto-classification and powerful taxonomy tools running
1. Manage your Group. 1. Log on to the CampusGroups platform.
1 1. Manage your Group 1. Log on to the CampusGroups platform. 2. Upon logging in, select your club from the Groups drop-down (next to the Home icon) and click on the Manage button next to the group of
[Ramit Solutions] www.ramitsolutions.com SEO SMO- SEM - PPC. [Internet / Online Marketing Concepts] SEO Training Concepts SEO TEAM Ramit Solutions
[Ramit Solutions] www.ramitsolutions.com SEO SMO- SEM - PPC [Internet / Online Marketing Concepts] SEO Training Concepts SEO TEAM Ramit Solutions [2014-2016] By Lathish Difference between Offline Marketing
Digital Marketing Training Institute
Our USP Live Training Expert Faculty Personalized Training Post Training Support Trusted Institute 5+ Years Experience Flexible Batches Certified Trainers Digital Marketing Training Institute Mumbai Branch:
Text Analytics Beginner s Guide. Extracting Meaning from Unstructured Data
Text Analytics Beginner s Guide Extracting Meaning from Unstructured Data Contents Text Analytics 3 Use Cases 7 Terms 9 Trends 14 Scenario 15 Resources 24 2 2013 Angoss Software Corporation. All rights
Evaluating Semantic Web Service Tools using the SEALS platform
Evaluating Semantic Web Service Tools using the SEALS platform Liliana Cabral 1, Ioan Toma 2 1 Knowledge Media Institute, The Open University, Milton Keynes, UK 2 STI Innsbruck, University of Innsbruck,
Using NVivo to Manage Qualitative Data. R e i d Roemmi c h R HS A s s e s s me n t Office A p r i l 6, 2 0 1 5
Using NVivo to Manage Qualitative Data R e i d Roemmi c h R HS A s s e s s me n t Office A p r i l 6, 2 0 1 5 Introductions Please share: Your name Department Position and brief description of what you
Annotea and Semantic Web Supported Collaboration
Annotea and Semantic Web Supported Collaboration Marja-Riitta Koivunen, Ph.D. Annotea project Abstract Like any other technology, the Semantic Web cannot succeed if the applications using it do not serve
DIIMS Records Classifier Guide
DIIMS Records Classifier Guide Featuring Content Server 10 Second Edition, November 2012 Table of Contents Contents 1. DIIMS Overview... 3 1.1 An Overview of DIIMS within the GNWT... 3 1.1.1 Purpose of
Draft Response for delivering DITA.xml.org DITAweb. Written by Mark Poston, Senior Technical Consultant, Mekon Ltd.
Draft Response for delivering DITA.xml.org DITAweb Written by Mark Poston, Senior Technical Consultant, Mekon Ltd. Contents Contents... 2 Background... 4 Introduction... 4 Mekon DITAweb... 5 Overview of
Flattening Enterprise Knowledge
Flattening Enterprise Knowledge Do you Control Your Content or Does Your Content Control You? 1 Executive Summary: Enterprise Content Management (ECM) is a common buzz term and every IT manager knows it
Technical Report. The KNIME Text Processing Feature:
Technical Report The KNIME Text Processing Feature: An Introduction Dr. Killian Thiel Dr. Michael Berthold [email protected] [email protected] Copyright 2012 by KNIME.com AG
GRAPHICAL USER INTERFACE, ACCESS, SEARCH AND REPORTING
MEDIA MONITORING AND ANALYSIS GRAPHICAL USER INTERFACE, ACCESS, SEARCH AND REPORTING Searchers Reporting Delivery (Player Selection) DATA PROCESSING AND CONTENT REPOSITORY ADMINISTRATION AND MANAGEMENT
So today we shall continue our discussion on the search engines and web crawlers. (Refer Slide Time: 01:02)
Internet Technology Prof. Indranil Sengupta Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur Lecture No #39 Search Engines and Web Crawler :: Part 2 So today we
Extend Table Lens for High-Dimensional Data Visualization and Classification Mining
Extend Table Lens for High-Dimensional Data Visualization and Classification Mining CPSC 533c, Information Visualization Course Project, Term 2 2003 Fengdong Du [email protected] University of British Columbia
DIGITAL MARKETING TRAINING
DIGITAL MARKETING TRAINING Digital Marketing Basics Keywords Research and Analysis Basics of advertising What is Digital Media? Digital Media Vs. Traditional Media Benefits of Digital marketing Latest
Web Mining. Margherita Berardi LACAM. Dipartimento di Informatica Università degli Studi di Bari [email protected]
Web Mining Margherita Berardi LACAM Dipartimento di Informatica Università degli Studi di Bari [email protected] Bari, 24 Aprile 2003 Overview Introduction Knowledge discovery from text (Web Content
Search Engine Marketing (SEM) with Google Adwords
Search Engine Marketing (SEM) with Google Adwords Account Setup A thorough account setup will ensure that your search engine marketing efforts are on a solid framework. This ensures the campaigns, ad groups
Zeeblio Reviewer's Guide
Zeeblio Reviewer's Guide Table of Contents Overview... 1 ABOUT THIS GUIDE... 1 HOW TO USE THIS GUIDE... 1 ABOUT LYCOS, INC.... 1 Introducing Zeeblio... 2 ZEEBLIO OVERVIEW... 2 ZEEBLIO MEDIA MANAGER...
Enhancing the relativity between Content, Title and Meta Tags Based on Term Frequency in Lexical and Semantic Aspects
Enhancing the relativity between Content, Title and Meta Tags Based on Term Frequency in Lexical and Semantic Aspects Mohammad Farahmand, Abu Bakar MD Sultan, Masrah Azrifah Azmi Murad, Fatimah Sidi [email protected]
CIRGIRDISCO at RepLab2014 Reputation Dimension Task: Using Wikipedia Graph Structure for Classifying the Reputation Dimension of a Tweet
CIRGIRDISCO at RepLab2014 Reputation Dimension Task: Using Wikipedia Graph Structure for Classifying the Reputation Dimension of a Tweet Muhammad Atif Qureshi 1,2, Arjumand Younus 1,2, Colm O Riordan 1,
ecommerce Web-Site Trust Assessment Framework Based on Web Mining Approach
ecommerce Web-Site Trust Assessment Framework Based on Web Mining Approach ecommerce Web-Site Trust Assessment Framework Based on Web Mining Approach Banatus Soiraya Faculty of Technology King Mongkut's
Enterprise 2.0 and SharePoint 2010
Enterprise 2.0 and SharePoint 2010 Doculabs has many clients that are investigating their options for deploying Enterprise 2.0 or social computing capabilities for their organizations. From a technology
NewsEdge.com User Guide
NewsEdge.com User Guide November 2013 Table of Contents Accessing NewsEdge.com... 5 NewsEdge.com: Front Page... 6 Saved Search View... 7 Free Text Search Box... 7 Company Watchlist... 9 Weather...12 NewsEdge.com:
Social Media Monitoring: Engage121
Social Media Monitoring: Engage121 User s Guide Engage121 is a comprehensive social media management application. The best way to build and manage your community of interest is by engaging with each person
WHAT'S NEW IN SHAREPOINT 2013 WEB CONTENT MANAGEMENT
CHAPTER 1 WHAT'S NEW IN SHAREPOINT 2013 WEB CONTENT MANAGEMENT SharePoint 2013 introduces new and improved features for web content management that simplify how we design Internet sites and enhance the
Web Document Clustering
Web Document Clustering Lab Project based on the MDL clustering suite http://www.cs.ccsu.edu/~markov/mdlclustering/ Zdravko Markov Computer Science Department Central Connecticut State University New Britain,
Terminology Extraction from Log Files
Terminology Extraction from Log Files Hassan Saneifar 1,2, Stéphane Bonniol 2, Anne Laurent 1, Pascal Poncelet 1, and Mathieu Roche 1 1 LIRMM - Université Montpellier 2 - CNRS 161 rue Ada, 34392 Montpellier
Bitrix Site Manager 4.1. User Guide
Bitrix Site Manager 4.1 User Guide 2 Contents REGISTRATION AND AUTHORISATION...3 SITE SECTIONS...5 Creating a section...6 Changing the section properties...8 SITE PAGES...9 Creating a page...10 Editing
Service Oriented Architecture
Service Oriented Architecture Charlie Abela Department of Artificial Intelligence [email protected] Last Lecture Web Ontology Language Problems? CSA 3210 Service Oriented Architecture 2 Lecture Outline
Single Property Website Quickstart Guide
Single Property Website Quickstart Guide Win More Listings. Attract More Buyers. Sell More Homes. TABLE OF CONTENTS Getting Started... 3 First Time Registration...3 Existing Account...6 Administration
Content Manager User Guide Information Technology Web Services
Content Manager User Guide Information Technology Web Services The login information in this guide is for training purposes only in a test environment. The login information will change and be redistributed
Developing a Collaborative MOOC Learning Environment utilizing Video Sharing with Discussion Summarization as Added-Value
, pp. 397-408 http://dx.doi.org/10.14257/ijmue.2014.9.11.38 Developing a Collaborative MOOC Learning Environment utilizing Video Sharing with Discussion Summarization as Added-Value Mohannad Al-Mousa 1
Elgg 1.8 Social Networking
Elgg 1.8 Social Networking Create, customize, and deploy your very networking site with Elgg own social Cash Costello PACKT PUBLISHING open source* community experience distilled - BIRMINGHAM MUMBAI Preface
Executive Dashboard Cookbook
Executive Dashboard Cookbook Rev: 2011-08-16 Sitecore CMS 6.5 Executive Dashboard Cookbook A Marketers Guide to the Executive Insight Dashboard Table of Contents Chapter 1 Introduction... 3 1.1 Overview...
Inner Classification of Clusters for Online News
Inner Classification of Clusters for Online News Harmandeep Kaur 1, Sheenam Malhotra 2 1 (Computer Science and Engineering Department, Shri Guru Granth Sahib World University Fatehgarh Sahib) 2 (Assistant
Taxonomy Enterprise System Search Makes Finding Files Easy
Taxonomy Enterprise System Search Makes Finding Files Easy 1 Your Regular Enterprise Search System Can be Improved by Integrating it With the Taxonomy Enterprise Search System Regular Enterprise Search
LinkZoo: A linked data platform for collaborative management of heterogeneous resources
LinkZoo: A linked data platform for collaborative management of heterogeneous resources Marios Meimaris, George Alexiou, George Papastefanatos Institute for the Management of Information Systems, Research
Knowledge Discovery using Text Mining: A Programmable Implementation on Information Extraction and Categorization
Knowledge Discovery using Text Mining: A Programmable Implementation on Information Extraction and Categorization Atika Mustafa, Ali Akbar, and Ahmer Sultan National University of Computer and Emerging
Nonprofit Technology Collaboration. Web Analytics
Web Analytics Contents What is Web Analytics?... 2 Why is Web Analytics Important?... 2 Google Analytics... 3 Using Major Metrics in Google Analytics... 6 Traffic Sources... 6 Visitor Loyalty... 9 Top
How to select the right Marketing Cloud Edition
How to select the right Marketing Cloud Edition Email, Mobile & Web Studios ith Salesforce Marketing Cloud, marketers have one platform to manage 1-to-1 customer journeys through the entire customer lifecycle
T HE I NFORMATION A RCHITECTURE G LOSSARY
T HE I NFORMATION A RCHITECTURE G LOSSARY B Y K AT H AGEDORN, ARGUS A SSOCIATES M ARCH 2000 I NTRODUCTION This glossary is intended to foster development of a shared vocabulary within the new and rapidly
Source Type Definitions
Source Type Definitions The Jobs2web Referral Engine sorts all of your visitor traffic into a Source Engine, and then categorizes it into a Source Type. Jobs2web has approximately 2,500 Source Engines
Best Practices for Managing Your Public Web Space and Private Work Spaces
Best Practices for Managing Your Public Web Space and Private Work Spaces So You re an Administrator to a Committee, Round Table, System User Group or Task Force? This Guide will introduce you to best
SEO and Internet Marketing. For Professionals
SEO and Internet Marketing For Professionals Effective from: Jan 2014 SEO TRAINING OUTLINE Internet AND Search Engine Basics What is Internet Marketing? Importance of Internet Marketing Types of Internet
Distributed Database for Environmental Data Integration
Distributed Database for Environmental Data Integration A. Amato', V. Di Lecce2, and V. Piuri 3 II Engineering Faculty of Politecnico di Bari - Italy 2 DIASS, Politecnico di Bari, Italy 3Dept Information
Project specification for suncoastreoholdings.com. Version 1.0
Project specification for suncoastreoholdings.com Version 1.0 This document lists the list of major functionality that this website will offer on its first version. 10/2/2012 Table of Contents Introduction...
Building a Database to Predict Customer Needs
INFORMATION TECHNOLOGY TopicalNet, Inc (formerly Continuum Software, Inc.) Building a Database to Predict Customer Needs Since the early 1990s, organizations have used data warehouses and data-mining tools
aloe-project.de White Paper ALOE White Paper - Martin Memmel
aloe-project.de White Paper Contact: Dr. Martin Memmel German Research Center for Artificial Intelligence DFKI GmbH Trippstadter Straße 122 67663 Kaiserslautern fon fax mail web +49-631-20575-1210 +49-631-20575-1030
Study Guide #2 for MKTG 469 Advertising Types of online advertising:
Study Guide #2 for MKTG 469 Advertising Types of online advertising: Display (banner) ads, Search ads Paid search, Ads on social networks, Mobile ads Direct response is growing faster, Not all ads are
Introduction. A. Bellaachia Page: 1
Introduction 1. Objectives... 3 2. What is Data Mining?... 4 3. Knowledge Discovery Process... 5 4. KD Process Example... 7 5. Typical Data Mining Architecture... 8 6. Database vs. Data Mining... 9 7.
1.00 ATHENS LOG IN 4 2.00 BROWSE JOURNALS 4
1.00 ATHENS LOG IN 4 1.01 What is Athens? 4 1.02 Accessing Cambridge Journals Online using Athens 4 2.00 BROWSE JOURNALS 4 2.01 Browse: By title 5 2.02 Browse: By subject 5 2.03 Browse: Favourites 5 2.04
Movie Classification Using k-means and Hierarchical Clustering
Movie Classification Using k-means and Hierarchical Clustering An analysis of clustering algorithms on movie scripts Dharak Shah DA-IICT, Gandhinagar Gujarat, India [email protected] Saheb Motiani
72. Ontology Driven Knowledge Discovery Process: a proposal to integrate Ontology Engineering and KDD
72. Ontology Driven Knowledge Discovery Process: a proposal to integrate Ontology Engineering and KDD Paulo Gottgtroy Auckland University of Technology [email protected] Abstract This paper is
Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information
Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information Satoshi Sekine Computer Science Department New York University [email protected] Kapil Dalwani Computer Science Department
Online Reputation Management Services
Online Reputation Management Services Potential customers change purchase decisions when they see bad reviews, posts and comments online which can spread in various channels such as in search engine results
IT Academy Lesson Plan
10 IT Academy Lesson Plan Microsoft Sharepoint Turn potential into success Microsoft Office SharePoint 2010: Lesson Plans Introduction Preparing to teach courses on Microsoft SharePoint 2010 for the first
Copyright EPiServer AB
Table of Contents 3 Table of Contents ABOUT THIS DOCUMENTATION 4 HOW TO ACCESS EPISERVER HELP SYSTEM 4 EXPECTED KNOWLEDGE 4 ONLINE COMMUNITY ON EPISERVER WORLD 4 COPYRIGHT NOTICE 4 EPISERVER ONLINECENTER
ONTOLOGY FOR MOBILE PHONE OPERATING SYSTEMS
ONTOLOGY FOR MOBILE PHONE OPERATING SYSTEMS Hasni Neji and Ridha Bouallegue Innov COM Lab, Higher School of Communications of Tunis, Sup Com University of Carthage, Tunis, Tunisia. Email: [email protected];
Dashboard Admin Guide
MadCap Software Dashboard Admin Guide Pulse Copyright 2014 MadCap Software. All rights reserved. Information in this document is subject to change without notice. The software described in this document
Legal Informatics Final Paper Submission Creating a Legal-Focused Search Engine I. BACKGROUND II. PROBLEM AND SOLUTION
Brian Lao - bjlao Karthik Jagadeesh - kjag Legal Informatics Final Paper Submission Creating a Legal-Focused Search Engine I. BACKGROUND There is a large need for improved access to legal help. For example,
Mining Text Data: An Introduction
Bölüm 10. Metin ve WEB Madenciliği http://ceng.gazi.edu.tr/~ozdemir Mining Text Data: An Introduction Data Mining / Knowledge Discovery Structured Data Multimedia Free Text Hypertext HomeLoan ( Frank Rizzo
GET OPTIMIZED! HOW TO OPTIMIZE YOUR SHOPPING CENTER IN THE ONLINE ENVIRONMENT
GET OPTIMIZED! HOW TO OPTIMIZE YOUR SHOPPING CENTER IN THE ONLINE ENVIRONMENT THINK DIFFERENTLY BUSINESS OBJECTIVE: GET MORE SHOPPERS TRADITIONAL STRATEGY EVENT OR PROMOTION RADIO PRINT DIRECT MAIL ONLINE
Bridging CAQDAS with text mining: Text analyst s toolbox for Big Data: Science in the Media Project
Bridging CAQDAS with text mining: Text analyst s toolbox for Big Data: Science in the Media Project Ahmet Suerdem Istanbul Bilgi University; LSE Methodology Dept. Science in the media project is funded
A QoS-Aware Web Service Selection Based on Clustering
International Journal of Scientific and Research Publications, Volume 4, Issue 2, February 2014 1 A QoS-Aware Web Service Selection Based on Clustering R.Karthiban PG scholar, Computer Science and Engineering,
Slide 7. Jashapara, Knowledge Management: An Integrated Approach, 2 nd Edition, Pearson Education Limited 2011. 7 Nisan 14 Pazartesi
WELCOME! WELCOME! Chapter 7 WELCOME! Chapter 7 WELCOME! Chapter 7 KNOWLEDGE MANAGEMENT TOOLS: WELCOME! Chapter 7 KNOWLEDGE MANAGEMENT TOOLS: Component Technologies LEARNING OBJECTIVES LEARNING OBJECTIVES
CINDOR Conceptual Interlingua Document Retrieval: TREC-8 Evaluation.
CINDOR Conceptual Interlingua Document Retrieval: TREC-8 Evaluation. Miguel Ruiz, Anne Diekema, Páraic Sheridan MNIS-TextWise Labs Dey Centennial Plaza 401 South Salina Street Syracuse, NY 13202 Abstract:
HydroDesktop Overview
HydroDesktop Overview 1. Initial Objectives HydroDesktop (formerly referred to as HIS Desktop) is a new component of the HIS project intended to address the problem of how to obtain, organize and manage
Taxonomies in Practice Welcome to the second decade of online taxonomy construction
Building a Taxonomy for Auto-classification by Wendi Pohs EDITOR S SUMMARY Taxonomies have expanded from browsing aids to the foundation for automatic classification. Early auto-classification methods
Computational Advertising Andrei Broder Yahoo! Research. SCECR, May 30, 2009
Computational Advertising Andrei Broder Yahoo! Research SCECR, May 30, 2009 Disclaimers This talk presents the opinions of the author. It does not necessarily reflect the views of Yahoo! Inc or any other
Job Streaming User Guide
Job Streaming User Guide By TOPS Software, LLC Clearwater, Florida Document History Version Edition Date Document Software Trademark Copyright First Edition 08 2006 TOPS JS AA 3.2.1 The names of actual
