ASWGF! Towards an Intelligent Solution for the Deep Semantic Web Problem



Similar documents
Proposed Protocol to Solve Discovering Hidden Web Hosts Problem

Chapter-1 : Introduction 1 CHAPTER - 1. Introduction

So today we shall continue our discussion on the search engines and web crawlers. (Refer Slide Time: 01:02)

Dr. Anuradha et al. / International Journal on Computer Science and Engineering (IJCSE)

DISCOVERING RESUME INFORMATION USING LINKED DATA

Semantic Search in Portals using Ontologies

A Knowledge Management Framework Using Business Intelligence Solutions

Web 3.0 image search: a World First

Unlocking The Value of the Deep Web. Harvesting Big Data that Google Doesn t Reach

Chapter 6. Attracting Buyers with Search, Semantic, and Recommendation Technology

Annotation: An Approach for Building Semantic Web Library

Lightweight Data Integration using the WebComposition Data Grid Service

ONTOLOGY-BASED APPROACH TO DEVELOPMENT OF ADJUSTABLE KNOWLEDGE INTERNET PORTAL FOR SUPPORT OF RESEARCH ACTIVITIY

Search and Information Retrieval

A Comparative Approach to Search Engine Ranking Strategies

HEALTH INFORMATION MANAGEMENT ON SEMANTIC WEB :(SEMANTIC HIM)

Arya Progen Technologies & Engineering India Pvt. Ltd.

A Semantic web approach for e-learning platforms

Get to Grips with SEO. Find out what really matters, what to do yourself and where you need professional help

Deriving Business Intelligence from Unstructured Data

City Data Pipeline. A System for Making Open Data Useful for Cities. stefan.bischof@tuwien.ac.at

Search Engine Optimization Content is Key. Emerald Web Sites-SEO 1

Supporting Change-Aware Semantic Web Services

SEO 360: The Essentials of Search Engine Optimization INTRODUCTION CONTENTS. By Chris Adams, Director of Online Marketing & Research

What do Big Data & HAVEn mean? Robert Lejnert HP Autonomy

EVILSEED: A Guided Approach to Finding Malicious Web Pages

Walk Before You Run. Prerequisites to Linked Data. Kenning Arlitsch Dean of the

EUR-Lex 2012 Data Extraction using Web Services

Web-Scale Extraction of Structured Data Michael J. Cafarella, Jayant Madhavan & Alon Halevy

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

Security Issues for the Semantic Web

Search Result Optimization using Annotators

Make search become the internal function of Internet

What to do Post Google Panda

Semantically Enhanced Web Personalization Approaches and Techniques

Search Engine Optimisation Guide May 2009

SEO Techniques for a Website and its Effectiveness in Context of Google Search Engine

A Framework for Developing the Web-based Data Integration Tool for Web-Oriented Data Warehousing

CHAPTER 1 INTRODUCTION

THE SEMANTIC WEB AND IT`S APPLICATIONS

Pragmatic Web 4.0. Towards an active and interactive Semantic Media Web. Fachtagung Semantische Technologien September 2013 HU Berlin

Design and Implementation of Domain based Semantic Hidden Web Crawler

Semantification of Query Interfaces to Improve Access to Deep Web Content

Introduction to Inbound Marketing

SEARCH ENGINE OPTIMIZATION

The Need for PDF Search Search and Index Overview IFilter Architecture Performance and Scalability Are Essential...

IBM Cognos Express Essential BI and planning for midsize companies

Structured Content: the Key to Agile. Web Experience Management. Introduction

An Ontology-based e-learning System for Network Security

Semantic Concept Based Retrieval of Software Bug Report with Feedback

Secure Semantic Web Service Using SAML

Performance Workload Design

Model-Driven Data Warehousing

SEO Definition. SEM Definition

Increasing Traffic to Your Website Through Search Engine Optimization (SEO) Techniques

Course Contents of Digital Marketing

ONTOLOGY FOR MOBILE PHONE OPERATING SYSTEMS

Search engine ranking

PUSH INTELLIGENCE. Bridging the Last Mile to Business Intelligence & Big Data Copyright Metric Insights, Inc.

Training Management System for Aircraft Engineering: indexing and retrieval of Corporate Learning Object

Published by: PIONEER RESEARCH & DEVELOPMENT GROUP ( 28

BUSINESS INTELLIGENCE AS SUPPORT TO KNOWLEDGE MANAGEMENT

Fig (1) (a) Server-side scripting with PHP. (b) Client-side scripting with JavaScript.

Well, it isn t if you have the right pre-built, totally unique website, which has all the income resources already built in.

A collaborative platform for knowledge management

Knowledge-Based Validation, Aggregation and Visualization of Meta-data: Analyzing a Web-Based Information System

Annotea and Semantic Web Supported Collaboration

Enhance Website Visibility through Implementing Improved On-page Search Engine Optimization techniques

ONTOLOGY-BASED MULTIMEDIA AUTHORING AND INTERFACING TOOLS 3 rd Hellenic Conference on Artificial Intelligence, Samos, Greece, 5-8 May 2004

Adding New Level in KDD to Make the Web Usage Mining More Efficient. Abstract. 1. Introduction [1]. 1/10

APPLICATION OF MULTI-AGENT SYSTEMS FOR NETWORK AND INFORMATION PROTECTION

An Intelligent Approach for Integrity of Heterogeneous and Distributed Databases Systems based on Mobile Agents

SEO: What is it and Why is it Important?

An Approach for Facilating Knowledge Data Warehouse

How To Make Sense Of Data With Altilia

A Framework of User-Driven Data Analytics in the Cloud for Course Management

Data Warehousing and OLAP Technology for Knowledge Discovery

Intinno: A Web Integrated Digital Library and Learning Content Management System

Google Analytics for Robust Website Analytics. Deepika Verma, Depanwita Seal, Atul Pandey

SEO Workshop Today s Coach Lynn Stevenson. SEO Analyst

SEO Basics for Starters

Building Semantic Content Management Framework

A HUMAN RESOURCE ONTOLOGY FOR RECRUITMENT PROCESS

What is a Mobile Responsive Website?

Standards, Tools and Web 2.0

Transcription:

ASWGF! Towards an Intelligent Solution for the Deep Semantic Web Problem Mohamed A. Khattab, Yasser Hassan and Mohamad Abo El Nasr * Department of Mathematics & Computer Science, Faculty of Science, Alexandria University, Egypt, * Head of Computer Engineering Department, Arab Academy for Science, Technology and Maritime, Egypt khattab@alex-sci.edu.eg, y.fouad@alex-sci.edu.eg, * m.aboelnasr@gmail.com Abstract Deep Semantic Web is a new era of Web problems correlated to the release of Semantic Web search engines. This research considers the problem of deep Semantic Web as a challenge for the future of the Web in order to achieve the dream of Semantic Web connecting everything. This paper addresses the problem of Semantic Web as being a solution that needs a solution and announces the term Deep Semantic Web in an attempt to motivate the researchers to take into consideration such an invisible problem and find a solution for it. Our contribution is trying to provide a state of the art solution through proposing an intelligent Semantic approach for Web search engines towards a homogeneous Semantic Web. The innovative idea of The Automatic Semantic Web Generator Framework ASWGF is embodied in the implementation and integration process with the current Web servers once. The approach considers the integration process as the spark off the model whereas the web servers will be the heart of the framework, we are not going to bet on the webmasters anymore in the success or the failure of the model. Index Terms Deep Semantic Web, Automatic Semantic Web Generator Framework, ASWGF, Semantic Web, Semantic Web crawler, Semantic search engine, Semantic Web approaches, Semantic Web publishing, Web crawler. 1. Introduction The largest known resource for information is the World Wide Web as people all over the world use the Web to find all sorts of information. The Web is inherently distributed and the data on the Web is stored in numerous sources and those sources have central points that provide search capabilities which are the Web search engines [1]. Traditionally, crawlers have only targeted a portion of the Web called the publicly indexable Web (PIW) [2]. This refers to the set of pages reachable purely by following hypertext links, ignoring the unknown domains and hosts which are called invisible Web. Jill Ellsworth [3] used the term invisible Web in 1994 to refer to the Web sites that are not registered in any search engine. Frank Garcia [4] used the term "invisible Web" in 1996 to refer to the Web sites that are possibly reasonably designed, but the designers did not bother to register them in any of the Web search engines; therefore, no one can find them. The first use of the specific term "Deep Web" occurred 2001 in a study by Michael Bergman [5] in which Bergman avoided the term "invisible Web". Over the years, search engines have discovered and learned how to track many factors to help them decide which pages are relevant to a given search. All such factors have to do with either where the word appears, or the structure of the Web. Search Engine Optimization (SEO) considers how search engines work, what people search for, the actual search terms typed into search engines and which search engines are preferred by the targeted audience. Optimizing Web sites may involve editing its content HTML and the associated coding to increase both its relevance to specific keywords and to remove barriers to the indexing activities of the search engines. Site owners recognize the value of having their Web sites highly ranked and visible in search engine results. Semantic Web crawler differs from a traditional Web crawler in the format of the source material it is traversing, and the means of specifying links between information resources [6]. This paper announces a new term; Deep Semantic Web which refers to Web sites that are traditionally optimized and registered through the Web search engines and will not be indexed through Semantic Web search engines; refers to the definition of term Deep Web. After the announcement 1

of Semantic Web search engines as alternatives to the traditional Web search engines, we can define the term Deep Semantic Web to refer to the surface Web from the point of view of Semantic Web search engines. Currently, there becomes a heterogeneity of structures between Web sites and Web search engines: as traditional HTML based Web sites, content management system (CMS) based Web sites optimized for traditional Web search engines or optimized for Semantic Web search engines, which Web sites being submitted to which Web search engine. Even Semantic based search engines and Semantic optimized Web sites use different approaches and ontologies which lead to the lack of optimization, and the Semantic dream is not achieved. 2. Intelligence in Semantic Web The Semantic Web is a mesh of information linked up in such a way as to be easily processed by machines on a global scale. Think of it as being an efficient way of representing data on the World Wide Web, or as a globally linked database; to do for data what the Web did for documents. Turning the Web into a database, and everyone's data becomes a part of it. The data becomes part of the worldwide database. The Semantic Web will give the ability to move from data record to data record, like going from Web page to Web page. Shortly, it is a service that aims at tieing it all together in order to make sense of the Web. The social graph is about connecting people while the Semantic Web is about connecting things. As illustrated in figure 1, it is not just connecting things but connecting everything so the intelligence is in the connections [7]. This makes the social graph just a small piece of the Semantic Web which revolves around connecting more kinds of things together. The connections will improve searching and advertising. Figure 1 The Semantic graph connects everything As the Web swells with more and more data, the predominant way of sifting through all of that data keyword search will one day break down in its ability to deliver the exact information. In fact, some argue that keyword search already delivers diminishing returns. However, anyone frustrated by the sense that it takes longer to find something on traditional Web search engines today than it did a year ago knows there is some truth in his argument. Keyword search is okay unless the information explosion continues. The amount of information on the Web keeps growing exponentially to accommodate all those seekers, and they feel compelled to put their own personal and social information onto the Web. At a certain point, with billions and billions of Web pages to sift through, just keyword search will not cut it anymore. Figure 2 illustrates the irrelative relation between the amount of searched data and the productivity of search across the Web generations. Figure 2 Irrelative relations between the amount of data and the productivity of search It is a needle in the haystack kind of problem. With the haystacks just getting bigger and bigger every second, keyword search engines return haystacks, but what we are truly looking for are the needles. The problem with the keyword search approach based is that only highly cited pages appear into the top results. The seeker gets a huge pile of results, but the page that the seeker wants the needle may not be highly cited by other pages and so it does not appear on the first page. This is because keyword search engines do not understand the search term; they just find pages that match the words searched for. What we really need is searching beyond the limits of keyword search. Semantic search is perhaps the most promising, but it essentially requires every single Web page to be re written. The future vision for the Web by 2

Tim Berners Lee inventor of the Web was the Semantic Web [8] to automate the process of collection and integration of information, but the dream is not yet realized. 3. Semantic Web Crawler vs. Web Crawler The Web crawler is a program or an automated script that browses the World Wide Web in a methodical and automated manner and is known as "Web spider" or Web robot". Other less frequently used names for Web crawlers are "ants", "automatic indexers", "bots", and "worms" [9]. The process of browsing the World Wide Web is called "Web crawling" or "spidering". Many sites, in particular search engines, use spidering as a means of providing up to date data. Web crawlers are mainly used to create a copy of all the visited pages for later processing by a search engine that will index the downloaded pages to provide fast searches. Crawlers can also be used for automating maintenance tasks on the Web sites, such as: checking links or validating HTML code. Also, crawlers can be used to gather specific types of information from Web pages, such as harvesting e mail addresses. In concept a Semantic Web crawler differs from a traditional Web crawler in only two regards: the format of the source material it is traversing, and the means of specifying links between information resources. Whereas a traditional Web crawler operates on HTML documents, linked using HTML anchors, a Semantic Web crawler scutter operates on RDF metadata with linking implemented using the rdfs. However, in practice, the aggregation and processing of Semantic Web content by a scutter differs significantly from that of a normal Web crawler. A normal crawler must only contend with extracting text from, possibly invalid, content and subsequent link extraction, whereas a Semantic Web crawler must carry out additional processing tasks: merging of information resources via inverse functional properties; tracking provenance of data; harvesting schemas and ontologies in addition to source data; extraction of embedded metadata (e.g. EXIF, XMP), etc [11] provides an excellent introduction to these issues and summarizes implementation experience gained whilst constructing a scutter. For example one of the Semantic Web crawlers is Slug [12], an open source, configurable and modular scutter. The emphasis of the framework is on modularity and ease of use rather than speed (hence the name!), although performance is currently acceptable. Figure 3 Typical high level architecture of a Web crawler Web crawlers are a central part of Web search engines, and details on their algorithms and architecture are kept as business secrets. When crawler designs are published, there is often an important lack of details that prevents others from reproducing the work. A Web crawler is one type of bot, or software agent and the typical high level architecture of Web crawlers is shown in figure 3 involving a scheduler and a multi threaded downloader [10]. Figure 4 Summary of slug architecture The UML diagram provided in figure 4 summarizes the key components and relationships present in the Slug framework. Central to the design of the framework are variations of the master slave and producer consumer design patterns. After presenting the two common Web crawler types, we need to think about the result of a Semantic Web crawler visiting a traditionally published 3

Web sites and a Web crawler visiting a Semantically published Web sites. 4. Approaches for Semantic Web Publishing Semantic publishing on the Web or Semantic Web publishing refers to publishing information on the Web as documents accompanied by Semantic markups [13]. Semantic publication is intended to provide a way for computers to understand the structure and even the meaning of the published information, making the search for information and the data integration more efficient. In the Semantic Web, information published on the Web is accompanied by metadata describing the published information; thus providing a Semantic context. There are many approaches for Semantic Web publishing that apply different trends and outputs to obtain a Semantic context as figure 5 shows. and author. The variety of approaches for Semantic Web publishing results in different outputs based on different ontology which lead to the inconsistency of Web contents and difficulty for Semantic Web search engines to index such contents. Figure 6 Generalized approach of dynamic Semantic publishing Figure 5 Approaches for Semantic Web publishing Data objects Publish information as data objects using Semantic Web languages like RDF [14] and OWL [15]. Embedded metadata Embed formal metadata in documents using new markup languages like RDFa [16] and Microformats [17]. Dynamic Semantic publishing A new technical term introduced by the online team of BBC [18]. It describes the idea of utilizing linked data technology to automate the aggregation and publication of interrelated content objects and provides rich content relationships as figure 6 illustrates a possible, generalized approach of dynamic Semantic publishing [19]. The content objects are multi dimensional entry points to a site and provide a much better and more user engaging way to explore content than the usual flat archives pages, which normally do not have dimensions beyond date, tag, 5. Semantic Web Search Engines Unlike traditional Web search engines, Semantic Web search engine stores Semantic information about Web resources and is able to solve complex queries, in addition to considering the context where the Web resource is targeted [20]. Using a Semantic Web search engine ensures more relevant results based on the ability to understand the definition of the word or the term that is being searched for, rather than on numbers. Semantic Web search engines are able to understand the context in which the words are being used, resulting in smart, relevant results which may seriously contribute to the development of electronic business applications since it is based on strong theory and widely accepted standards. After the announcement of releasing Semantic Web search engines as alternatives to traditional Web search engines, there is a list of the top seven Semantic Web search engines with a short description about each one to get started via Semantic Web searching [21]. The release of such Semantic Web search engines results in a heterogeneous Web contents composed of different architectural-based Web sites HTML/code-based or CMS-based, and different knowledge-based Web search engines traditional or Semantic in addition to different ontologies and methodologies even in the same type according to approaches for Semantic Web publishing as illustrated in figure 7. 4

is concerned with the automation of Semantic annotation process through the implementation of the proposed approach as an Automatic Semantic Web Generator Framework ASWGF. The third aspect is time-saving by involving the Web servers in the Semantic annotation process and guaranteeing the perfection of such process according to the proposed approach and the implemented framework. Figure 7 Heterogeneous Web contents The search engine optimization process for newly designed or developed Web sites to be submitted or indexed through a Web search engine should be done for traditional Web search engines or for Semantic Web search engines. If the Semantic will be the target, which ontology should be utilized to be indexed as there are different applied ontologies? And which type of Web crawler will traverse and index the contents of such Web sites? And how to do so? The heterogeneity of the Web contents leads to one or all of the following: Lack of optimization. Heterogeneous Web content. Inconsistence Semantic Web model. Deep Semantic Web. The Semantic Web results in a solution needs for a solution. 6. The Overview of the Proposed Approach The problem of deep Semantic Web is considered a challenge for the future of the Web in order to achieve the dream of Semantic Web connecting everything. We are trying here to provide a state of the art solution by proposing an intelligent Semantic approach for the Web search engines towards a homogeneous Semantic Web. 6.1 The Aspects of the Proposed Approach The proposed approach will maintain three aspects as illustrated in figure 8. The first aspect considers ontology unification from the side of the Web search engine to guarantee the speed and the simplicity of the indexing process for Web sites. Such aspect will guarantee the homogeneity of Web sites structures according to a pre defined ontology. The Web search engine will no longer need to analyze the contents of the Web site; thus saves the time for more enhanced processes. The second aspect Figure 8 Aspects of the proposed model 6.2 The Automatic Semantic Web Generator Framework ASWGF Webmasters as long as we are talking about Web problems or Web solutions always there is a spot on webmasters as they must be involved in any keywordbased optimization process or Semantic annotation process which always lead to the lack of solution or burden the Web search engines as must be the source and the provider for the solution. The innovative idea of The Automatic Semantic Web Generator Framework ASWGF is embodied in the implementation and integration process with the current Web servers once. The approach considers the integration process as the spark off the model whereas the web servers will be the heart of the framework, we are not going to bet on the webmasters anymore in the success or the failure of the model. The start of any Web chain is by registering a domain name then hosting a certain Web site over such registered domain and our Semantic Web generator claims such a starting point and operates through a sequence of processing phases as follows: Phase i (crawling & analyzing) based on the hosting Web server and by applying the Host List Protocol model [22] listing the hosted domains, crawling and analyzing 5

their contents regardless to the content type HTML based or CMS based. Phase ii (decision & Semantic annotation) after the analysis process the model based on reasoning will specify the target domain and decide which Semantic repository is suitable for such crawled and analyzed content to use it and complete the Semantic annotation process. Phase iii (Semantic Web generation) the avenues during this phase may be in one of two directions: 1. Send the resulting Semantic content to the Web search engines to handle the received information according to the pre-agreed ontology. 2. Building an intermediate repository (core of new Semantic Web search engine) to be accessed through end users / search engines according to the policy agreed upon. 7. Experimental Results One of the objectives of our model is to unify the processing phases regardless to the underlying system of the Website content HTML-based, CMS-based or other Web script-based. The experiment demonstrates the capability of the framework extracting such heterogeneous content during preserving the reliability of the system according to the accuracy rate factor. The environment for the experiment was a random sample of the World Wide Web hosted Websites classified into four main categories according to their platform including the operating system, the underlying server daemons and the development environment which was an average of five values for each environment. Accuracy rate and error rate performed on the data series obtained through the experiment. 1 0.8 0.6 0.4 0.2 HTML CMS PHP ASPX 0 1 2 3 4 5 Figure 10 Distribution of the results accuracy relative to the underlying development environment Figure 9 Members of the ASWGF Framework The state of the art in the proposed approach can be summarized as follows: Automating the Semantic Web generation process, not just the Semantic annotation process [23]. Unifying the processing phases regardless to the underlying system of the Website content HTML-based CMS-based or other Web script-based. Minimizing the processing time single processing time as all Web servers will work at the same time instead of sending crawlers of Web search engines to crawl the Web site contents searching for traditionally or Semantically optimized content. According to figure 10 the best accuracy value is ranging from 98% to 100% for all underlying development environment while the worst accuracy value is ranging from 75% to 84% for ASPX underlying development environment and such for two cases of the five averages. The evaluation of the obtained accuracy rates performed on the data series obtained through the experiment leads to the fact that the framework successfully extracted such heterogeneous content during preserving the reliability of the system in terms of the accuracy rate. 8. Conclusion and Future Work Semantic Web is a solution that needs a solution. The announcement of Semantic Web search engines as alternatives to traditional Web search engines affects the surface Web as moving slowly towards being deep Semantic Web. The release of Semantic Web results in a heterogeneous Web contents composed of different architectures based Web sites HTML/code-based or 6

CMS-based and different knowledge based Web search engines traditional or Semantic in addition to different ontologies and applied approaches. The problem of Web heterogeneity is a challenge and a threat to the future of the Web as the Semantic Web is to the future. The heterogeneity of Web contents leads to the lack of optimization, heterogeneous Web, inconsistent Semantic Web model and deep Semantic Web. We are trying here to provide a state of the art solution through proposing an intelligent based approach results in an automatic Semantic Web Generator Framework ASWGF towards a homogenous Semantic Web which is the vision of the future for the Web. The evaluation of the obtained accuracy rates performed on the data series obtained through the experiment leads to the fact that the framework successfully extracted such heterogeneous content during preserving the reliability of the system in terms of the accuracy rate. References [1] S. Lam, The Overview of Web Search Engines, University of Waterloo, Canada, February 2001. [2] S. Lawrence and C. L. Giles, Searching the World Wide Web, Web Science magazine, 280(5360):98, 1998. [3] Jill H. Ellsworth, Matthew V. Ellsworth, The internet business book, John Wiley & sons, Inc., New York, NY, 1994. [4] Garcia, Frank., Business and marketing on the Internet, Masthead 9 (1), 2001. [5] Bergman, Michael K, The deep Web: surfacing hidden value. The journal of electronic publishing, 7 (1), Aug 2001. [6] Dan Brickley. RDF Hyper linking, http://www.w3.org/2001/sw/europe/talks/xml 2003/Overview 6.html, 2003. [7] Nova Spivack, The Next Web Conference, Amsterdam, Netherlands, April 2008. [8] Tim Berners Lee, Invents the Web Again, http://www.ted.com/talks/tim_berners_lee_on _the_next_web.html, retrieved on 2011. [9] Kobayashi, M. and Takeda, K, Information retrieval on the Web. ACM Computing Surveys (ACM Press), 32 (2): 144 173. doi:10.1145/358923.358934, 2000. [10] Carlos Castillo, Effective Web crawling, ACM SIGIR Forum, v.39 n.1, June 2005. [11] Matt Biddulph, Crawling the Semantic Web, BBC London, United Kingdom, June 2003. [12] Leigh Dodds, Slug: A Semantic Web Crawler, February 2006. [13] Semantic publishing, http://en.wikipedia.org/wiki/semantic_publish ing, retrieved on 2011. [14] RDF Primer, W3C Recommendation http://www.w3.org/tr/rdf primer, February 2004. [15] OWL Web Ontology Language Reference, W3C Recommendation, http://www.w3.org/tr/owl ref, February 2004. [16] RDFa Primer, Bridging the Human and Data Webs W3C Working Draft, http://www.w3.org/tr/2008/wd xhtml rdfa primer 20080620, June 2008. [17] W3C GRDDL Recommendation Bridges HTML/Microformats and the Semantic Web, http://xml.coverpages.org/ni2007 09 13 a.html, September 2007. [18] BBC World Cup 2010 dynamic Semantic publishing, http://bnode.org/blog/2010/07/30/dynamic Semantic publishing for any blog part 1, July 2010.. [19] Dynamic Semantic Publishing for any Blog, http://bnode.org/blog/2010/07/30/dynamic Semantic publishing for any blog part 1, July 2010. [20] Alcides Calsavara and Glauco Schmidt, ADVANCED DISTRIBUTED SYSTEMS, Lecture Notes in Computer Science, Volume 3061/2004, 145 157, DOI: 10.1007/978 3 540 25958 9_14, 2004. [21] Top 7 Semantic Search Engines As An Alternative To Google, http://www.makeuseof.com/tag/top 7 Semantic search engines alternative google search, retrieved on 2011. [22] Kahttab, M. A., Y. Fouad and O. A. Rawash, Proposed Protocol to Solve Discovering Hidden Web Hosts Problem, IJCSNS, Vol. 9 No. 8 pp. 247 254, 2009. [23] Innovations in Computing Sciences and Software Engineering, 523 525, DOI: 10.1007/978 90 481 9112 3_89, 2010. 7