Towards spatial data quality information analysis tools for experts assessing the fitness for use of spatial data

Size: px
Start display at page:

Download "Towards spatial data quality information analysis tools for experts assessing the fitness for use of spatial data"

Transcription

1 International Journal of Geographical Information Science Vol. 21, No. 3, March 2007, Research Article Towards spatial data quality information analysis tools for experts assessing the fitness for use of spatial data R. DEVILLERS*{{, Y. BÉDARD{, R. JEANSOULIN{ and B. MOULIN{ {Department of Geography, Memorial University of Newfoundland, St. John s (NL) A1B 3X9, Canada {Centre de Recherche en Géomatique (CRG), Pavillon Casault, Université Laval, Québec (QC), G1K 7P4, Canada Laboratoire des Sciences de l Information et des Systèmes (LSIS), Centre de Mathématiques et d Informatique (CMI), Université de Provence (Aix-Marseille I), 39 Rue Joliot Curie, Marseille Cedex 13, France (Received 4 October 2004; in final form 11 November 2005 ) Geospatial data users increasingly face the need to assess how datasets fit an intended use. However, information describing data quality is typically difficult to access and understand. Therefore, data quality is often neglected by users, leading to risks of misuse. Understanding data quality is a complex task that may involve thousands of partially related metadata. For complex cases where heterogeneous datasets have to be integrated, there is a need for tools supporting data quality analysis. This paper presents the design of such a tool that can manage heterogeneous data quality information and provide functions to support expert users in the assessment of the fitness for use of a given dataset. Combining concepts from GIS and Business Intelligence, this approach provides interactive, multi-granularity and context-sensitive spatial data quality indicators that help experts to build and justify their opinions. A prototype called the Multidimensional User Manual is presented to illustrate this approach. Keywords: Spatial data quality; Fitness for use; Visualization; Indicators; Spatial OLAP; Dashboard; Metadata 1. Introduction The last decade has witnessed a major trend towards the democratization of geospatial data. A recent example is Google Earth, which is becoming increasingly popular among the general public as well as being used increasingly by specialists to develop new applications. Geospatial data are now used in various application domains and by a variety of users, from experts using highly sophisticated systems to mass-users using web or mobile mapping technologies. In spite of being a positive evolution, such democratization also facilitates the use of data for non-intended purposes, as well as the overlaying of heterogeneous data collected at different periods, by different organizations, using various acquisition technologies, standards, and specifications. Such a context increases the risks of geospatial data misuse. In this sense, Goodchild (1995) argues that GIS is its own worst enemy: by inviting people to find new uses for data, it also invites them to be irresponsible in *Corresponding author. rdevillers@mun.ca International Journal of Geographical Information Science ISSN print/issn online # 2007 Taylor & Francis DOI: /

2 262 R. Devillers et al. their use. Such cases of data misuse have already occurred, sometimes leading to significant social, political, or economical impacts (see, for instance, Beard 1989, Monmonier 1994, Agumya and Hunter 1997 and Gervais 2003, for various examples of cases of data misuse). One solution is to describe more explicitly users needs and to better assess the fitness of certain data for a specific use over a given area. This remains, however, a very difficult task, among other reasons, due to the inherent complexity of the problem and to the typically inadequate documentation regarding data specifications (in spite of the development of metadata standards over the past 10 years such as FGDC, ANZLIC, CEN, and, more recently, ISO). Efforts have been deployed to encourage better use of metadata, especially for National Spatial Data Infrastructures (NSDI), while at the same time, an increasing number of papers have been published regarding the evaluation of the fitness for use (e.g. Frank 1998, Agumya and Hunter 1999a, b, De Bruin et al. 2001, Vasseur et al. 2003, Frank et al. 2004, Grum and Vasseur 2004). However, assessing the fitness for use remains a task that rapidly becomes very complex, especially when the data for the area of interest come from sources that are heterogeneous spatially, temporally, and, even more problematic, semantically. Consequently, more fundamental research is needed to provide better quality assessment methods, notwithstanding better tools to facilitate this task. From a legal perspective, as geospatial data become a mass-product, considerations about consumer protection, liability, clear instruction manuals with warnings, guarantees and similar issues arise. In this context, meaningful quality information must be available to help users or providers to assess the fitness of the data for a given purpose (Gervais 2003, 2006). So far, this has barely succeeded. Although metadata (i.e. data about data) currently distributed by data producers should help the users in this task, from the end-user standpoint, they are typically expressed in a cryptic language, recorded using a complex structure, lacking explicit links with the data they describe, or simply inexistent in many cases. Hence, data-quality information is not easily accessible, understandable, or adapted to the usage context and users need. This could be regarded as a fault from a legal point of view (Gervais 2006). In the context of our research, we consider that experts familiar with spatial data quality issues should get involved in the process (figure 1), either internally within the organization or through mandates with external experts involving their professional liability (Gervais 2003, Bédard et al. 2004, Devillers 2004). Similar scenarios have already been going for some time (e.g. Gervais et al. 2006, Marttinen 2006, Smith 2006) and are becoming more frequent within organizations involved in data-warehousing initiatives. Figure 1. Quality information system objective.

3 Quality information analysis tool 263 In order to support these quality experts in the assessment of the fitness for use, there is a need to provide a more effective approach to express quality information than the simple distribution of metadata. Improved methods and tools would facilitate the integration, management, and communication of information about data quality, allowing experts to increase their knowledge about data quality, and to better assess how data fit an intended use. Several authors have recently mentioned the need for such methods and tools. For instance, Lowell (2004) expresses the need for a computer-based intelligent engine that could analyse information about uncertainty. He argues that: Humans will not be able to absorb and assimilate all of the information presented in an uncertainty-based database, and will not have the capacity to analyse all of it efficiently. This will require the creation of new analytical and visualization tools capable of providing humans with a logical summary of the uncertainty information present in the system. Due to the complexity of this task, we think that it is currently impossible to design a system providing a clear output regarding the fitness or not of the data for a certain use. We argue that the only possibility available today and in the near future is to provide quality experts with the available information regarding data quality, either at the detailed level of existing metadata or at more general levels resulting from metadata aggregation/generalization (when meaningful) or past quality analysis. It is our belief that no system using fully automatic methods will provide clear answers to such questions, as these are more than technological issues and require in-depth personal understanding of the data that can only be answered by a professional expert. Consequently, the objective of this paper is to present a Spatial data quality Information Analysis tool called the Multidimensional User Manual (MUM) that aims at supporting experts in spatial data quality (rather than supporting the end user of the data) in an innovative manner. This solution makes use of quality indicators organized hierarchically and uses a combination of Spatial On-Line Analytical Processing (SOLAP) and Dashboard tools. The proposed solution is intended to contribute to the professional activity of Quality Reporting and Quality Auditing. In the present paper, we first discuss the issues of managing and communicating spatial data quality information. Then, we describe our approach that uses quality indicators that are based on quality information stored in a multidimensional datacube (in OLAP vocabulary). We name this structure the Quality Information Management Model (QIMM). The advantages of using indicators are explained, as well as their limitations. Top-down and bottom-up approaches, two complementary ways to populate the QIMM, are then discussed. We then present our MUM prototype which supports several techniques to manage, interactively explore, and communicate quality information at different levels of detail to expert users and data quality experts. The use of Spatial On-Line Analytical Processing (SOLAP) functions, as well as the general architecture of the prototype, is described and discussed. This includes the different functions available and how such approaches can facilitate the quality-assessment process (SOLAP fundamentals can be found in Rivest et al. 2001, 2005, and Bédard et al. 2003, in press). Finally, we discuss the integration of MUM into the professional activities of Quality Reporting and Quality Auditing and present future research directions.

4 264 R. Devillers et al. 2. Geospatial-data-quality management and communication For about 30 years, two different meanings have been associated with the term data quality in the literature, the first restricting quality to the absence of errors in the data (i.e. internal quality) and the second looking at how data fit the user s needs (i.e. external quality; Juran et al. 1974, Morrison 1995, Aalders and Morrison 1998, Aalders 2002, Dassonville et al. 2002, Devillers and Jeansoulin 2006). This second definition, usually identified as the concept of fitness for use (Juran et al. 1974, Chrisman 1983, Veregin 1999), is the one that reached an official agreement by standardization bodies (e.g. ISO) and international organizations (e.g. IEEE). More accurately, for the latter case, we define quality as the closeness of the agreement between data characteristics and the explicit and/or implicit needs of a user for a given application in a given area. For more than 20 years, standardization bodies have identified characteristics describing internal quality (e.g. ICA, FGDC, CEN, ISO). Despite some differences between these standards regarding the characteristics identified, there is a general agreement among most of them. The common criteria are often identified as the famous five : positional accuracy, attribute accuracy, temporal accuracy, logical consistency, and completeness (Guptill and Morrison 1995, ISO/TC ). From the end-user perspective, knowledge about internal data quality typically comes from the metadata transmitted with datasets by data producers. One metadata main objective is to allow end users to assess the fitness of a dataset for their use (ISO/TC ). However, academic studies and practical experience clearly show the limited benefit of metadata in their current form (Timpf et al. 1996, Frank 1998, Fisher 2003, Gervais 2003). End users frequently do not use metadata; indeed, it is not uncommon for users to ask producers not to include the metadata when ordering data. However, this is not an indication that metadata are useless, but an indication that metadata representation does not suit user needs. Our experience is that users rarely utilize metadata beyond those necessary for ordering datasets from digital libraries (e.g. searching datasets based on spatial and temporal extents or keywords). In addition to the inadequate form of today s metadata, which is typically too cryptic for most users, metadata are often too general to enable an adequate assessment of quality. This hides most of the information richness which should be communicated. Hunter (2001) clearly illustrates this point by giving several examples of existing metadata such as Positional Accuracy being variable, 100 m to 1000 m or 1.5 m (urban) to 250 m (rural). Such metadata rapidly become useless when someone wants to know, for instance, the quality of data for a certain region, object class (e.g. buildings) or object instance (e.g. one specific building). Despite higher costs, some data producers already provide certain metadata at the feature instance level (e.g. NTDB product of Geomatics Canada, LCM2000 of the Centre for Ecology and Hydrology, MasterMap of the Ordnance Survey). The recent international standard ISO (ISO/TC ) also supports metadata encoding down to the feature instance and attribute instance levels, while this was not possible with older metadata standards such as FGDC. Moreover, today s metadata are typically provided in a different file than the data, reducing the possibility of easily exploiting quality information directly from certain GIS functions. Consider the simple example of a distance measurement between two objects. Today s typical GIS provide answers that are related to the encoding of the data (e.g. integer, double) and not the accuracy of the objects that

5 may be documented in the metadata (e.g. ArcGIS provides distances with six decimals, i.e. a spatial precision of 1 mm, corresponding to the average diameter of a bacterium). Given the appropriate metadata, it would be possible to calculate automatically the accuracy of the distance measured and display the result accordingly (e.g. by providing error margins with the distance or just informing the user of the level of uncertainty/reliability related to this measure). Hence, it is possible to benefit from the quality information described in metadata. These benefits would be twofold: 1. A more efficient communication and assessment of quality information (an issue discussed in this paper) would help users to understand the limitations of the data. 2. The management of quality information within a structured database, when associated with a GIS tool, would provide results adapted to the data manipulated for the area of interest (this is a research perspective). Both points would help to reduce the risk of misuse and then reduce the occurrence of adverse consequences. Over the last 20 years, several research projects have focused on ways to better describe and communicate quality/uncertainty/error information. Proposed solutions were based on either legal and procedural methods (Bédard 1987, Agumya and Hunter 1999b, Gervais 2003, 2006) or visualization techniques (Buttenfield and Beard 1991, 1994, Beard and Mackaness 1993, Buttenfield 1993, McGranaghan 1993, Fisher 1994, Howard and MacEachren 1996, Beard 1997, Beard and Buttenfield 1999, Leitner and Buttenfield 2000, Drecki 2002, Devillers and Jeansoulin 2006), and on the communication of visual or audio warnings to users (Fisher 1994, Hunter and Reinke 2000, Reinke and Hunter 2002). However, none of these two techniques is yet implemented within commercial GIS software. Furthermore, none of these solutions allows users to navigate easily and rapidly from one quality characteristic to another, from one level of detail to another, from one area to another, from one category of data to another, from one feature to another, and so on, in order to help decide whether or not a dataset would fit the expressed needs for the area and objects of interest. Such functionalities are characteristic of the analytical data structure typical of modern decision-support technologies based on datacubes, such as On-Line Analytical Processing (OLAP), dashboards, datamarts and data mining. 3. Quality indicators and Quality Information Management Model (QIMM) 3.1 Quality indicators Quality information analysis tool 265 Given that spatial data quality information can be described using many different characteristics (e.g. horizontal and vertical accuracy, omission, commission, topological consistency, up-to-dateness), and given that data producers and standards tend towards a description of feature-level metadata, the volume of quality information increasingly becomes a problem when trying to convey this information efficiently. In many fields, people have to cope with similar problems of meaningfully communicating large volumes of data to support decision-making processes. They often use indicators (also named indices or index depending on the field) that can be displayed on dashboards (also named balanced scorecards or executive dashboards ) to communicate relevant information to decision-makers

6 266 R. Devillers et al. (Kaplan and Norton 1992, Fernandez 2000, von Schirnding 2000, Goglin 2001). We have adapted this traditional indicator-based approach for spatial data quality communication (Devillers et al. 2004) and implemented the resulting approach in a spatial datacube. Indicators can be defined as a way of seeing the big picture by looking at a small piece of it (Plan Canada 1999). Fernandez (2000) defines indicators as information or a group of information helping the decision-maker to appreciate a situation. Indicators show what is going on globally, allowing or not examination of details. Let us make an analogy to a family doctor who wants to diagnose his patient s illness. The doctor knows that the human body is a complex system and that he cannot observe and measure all of its characteristics. Hence, he uses certain observations and measures (e.g. body temperature, blood pressure, pulse) to gain a broad view of the patient s health. If one of these indicators shows a problem, the doctor can then use more advanced tests to investigate the potential problem into more details (e.g. blood analysis, X-ray). In similar ways, a number of organizations use indicators to assess larger complex systems (e.g. economic indicators, social indicators, or environmental indicators). Klein (1999) observed different types of decision-makers that have to make rapid and important decisions (e.g. firemen, aircraft pilots, surgeons) and, based on these observations, built the Recognition- Primed Decision model that is well known in the decision-making community. Klein argues that decisions of any type are driven by a combination of parameters, including the similarity the situation has with previous situations experienced by the person, the person s intuition and capacity to imagine potential scenarii, etc. His thesis confirms the importance of experts as a support in complex decision-making processes. He observed that indicators, named cues in his model, are key components in decision-making processes and are used to characterize situations and choose which action to perform. Indicators are seen as synthetic key information about complex phenomena providing global pictures and major trends. Several indicators can be aggregated in higher-level indicators, sometimes named aggregate indicators (or indices ). Typical strategic decision-making processes use a small number of indicators, or aggregate indicators, as one may see in numerous Business Intelligence (BI) applications and Executive Information Systems (EIS). Typical indicators can be drilled down in a small number of layers that are expanded to provide available details when needed. Selecting the most relevant indicators among available ones or collecting new data to build a new indicator represents an interesting challenge when designing decision-support systems. Using indicators for quality assessment appears not only theoretically interesting, but realistically unavoidable in order to build a usable and credible system. Some could say well it s not perfect but that s the best we can do, while others could be happier: well it is sufficient to make my decision! However, if indicators and aggregate indicators appear to be unavoidable when one wants to cope with large volumes of information, summarizing different data into a single value is a difficult task that has been discussed and questioned by many authors (e.g. Ott 1978, Meadows 1998, Jollands et al. 2003). Hence, the scientific community is still divided between different types of approaches, such as providing aggregate indicators or indicator profiles (matrices). In the first case, the aggregation process is carried out using a mathematical equation (e.g. aggregation of indicators using their average value), whereas in the second case, the decision-maker will do this process cognitively. Both approaches have positive and negative aspects,

7 Quality information analysis tool 267 and there is no perfect solution. The most important problem with aggregate indicators is that it requires many assumptions and decisions about the user needs (e.g. selecting an aggregation method and eventually weights for the indicators). Another important issue is to leave the possibility to the decision-maker to access the original data if needed. For applications requiring too many indicators, as is the case with spatial data quality (see, for instance, the different quality sub-elements suggested by the ISO/TC211 standards), the objective is then to minimize the disadvantages of the aggregate indicator approach. With this in mind, context-sensitive quality information can be provided to the user at the right level of abstraction in order to help them identify aspects of quality which are relevant for the task at hand. Furthermore, it is not uncommon to see metadata about indicators built into SOLAP applications as well as the access to the finest granularity of data and the possibility of selecting between aggregation methods or adding a new one on the fly. To analyze the fitness for use of geospatial data for a given area, we designed the MUM analysis tool such that quality indicators would be displayed on a dashboard embedded within a cartographic interface (i.e. SOLAP), acting as a quality information tool that could support quality experts in the assessment of the fitness of datasets of a specific use. In order to support the quality experts when investigating the robustness of the data obtained (i.e. something every professional must do when building their opinion), the characteristics of our approach include, among others: N providing a limited number of indicators and aggregate indicators to the user; N allowing the user to access the initial detailed data used to calculate the indicators and aggregate indicators; N allowing the user to select an aggregation method that best fits his needs, thus not automatically taking this decision for them. The aggregation method will for instance depend on the level of risk the user may accept. For example, a user who does not want to take any risk may use an aggregation technique that will propagate a problem from a detailed level (e.g. poor spatial accuracy of a specific road) to an aggregated indicator (e.g. spatial accuracy of the whole dataset). Another user more tolerant to the risk may average indicators values to obtain an aggregated indicator that would not highlight exceptional cases of low or high quality at the detail level. With some OLAP and SOLAP tools, it is even possible to add new indicators and calculate them on the fly for every level of aggregation; some also support operators that go beyond the typical sum, count, min, and max, supporting operators such as log, power, ln, trigonometric functions, or any other type of aggregation that could be made explicit; N allowing the user to set their own weights for their indicators, depending on their context; N providing the possibility to visualize any data quality indicator on a map at the most detailed level (i.e. feature instance), then allowing the possibility of pinpointing potential problems that would not be visible at aggregated levels; N providing access to the initial metadata used to derive the indicators. This whole approach relies on the assumption that the system should not take the decision for the expert and should not hide variations invisible from indicators. Rather, this approach is context-dependent. It aims at providing the expert with meaningful information about spatial data quality to support the quality analysis

8 268 R. Devillers et al. process, whatever the levels of detail and coverages needed in space, time, feature, aggregation method, etc. It is the expert s responsibility to choose and feed the indicators, and to explore this information to make their opinion. Therefore, the expert can produce a Quality Report or Audit Report with the quantity and level of information required, including a meaningful set of indicators for the task at hand. Each quality indicator can be based on one or several raw metadata. Metadata can be obtained from the traditional metadata provided with the datasets (e.g. metadata describing data quality or other metadata). These typical metadata can follow metadata standards (e.g. ISO/TC 211). But metadata used by MUM are not restricted to these typical metadata and can also come from other sources of information describing data quality, such as an expert opinion, an organization s internal consensus (e.g. lowest spatial precision for a given area or a lowest degree of completeness for a certain period within a dataset), or information about data quality derived from other calculations relevant to the user. Metadata are then not restricted to those described by the existing metadata standards but more generally refer to any data about data. 3.2 Quality Information Management Model (QIMM) A central motivation in this research is to avoid an information overload to users (i.e. the quality experts) when they have to access, compare, and analyse various data quality characteristics at different levels of detail, for different regions, different data sources, etc. According to the well-known psychological research from Miller (1956), human short-term memory (or working memory) can deal with only five to nine chunks of information at once. Hence, it would be of limited use to communicate a larger quantity of information simultaneously to a user. In addition, other psychological studies have shown that the length of time information stays in short-term memory (STM) is very limited (Baddeley 1997). This duration can be quite variable depending on the modality (i.e. acoustic, visual, or semantic), the necessity of performing actions (e.g. selecting an item on the screen of a computer), and other factors (for instance, the level of concentration). Experimental results usually provide durations varying from 2 to 30 s. According to Newell s (1990) physical and biological tests, among the four computational bands emerging from the natural hierarchy of information processing, response times between and 10 1 seconds are needed to perform cognitive tasks and maintain a line of thought. Consequently, an efficient method to convey quality information should limit the volume of information to less than nine chunks and provide information to users in less than 10 s in order to avoid interrupting their mind-stream. Another point highlighted by Reinke and Hunter (2002) is the need not only to passively obtain quality information from the system, but also to be able to interact with the system to request additional information (i.e. feedback loop). To cope with all these constraints, we based our approach on the multidimensional database models (datacube) used in the field of Business Intelligence (e.g. data warehousing, OLAP, data mining). In this field, multidimensional refers not only to spatial (x, y, z) and temporal (t) dimensions, as in the GIS domain, but also to semantic, temporal, and spatial hierarchies of concepts called dimensions, which are represented by the metaphor of a data hypercube containing data units (facts); each fact contains measures resulting from the intersection of all dimensions at a given level in their respective hierarchies (e.g. Berson and Smith 1997). For instance, a spatial dimension in Spatial OLAP will then not be a spatial axis on

9 Quality information analysis tool 269 which a measure can be done, as it is in GIS, but a hierarchy of spatial objects representing different levels of granularity (e.g. continentrcountryrregion/state; object classrobject instancergeometric primitive). Multidimensional database approaches appeared in the early 1980s (Rafanelli 2003), and numerous books and papers have been published on this vast topic, especially after it became popular in the mid-1990s. Codd (1993) clearly explained their superiority over relational databases when users need to interactively analyse large volumes of data. Datacubes provide an intuitive and very fast data structure supporting interactive analyses at the speed of thought when using OLAP tools (Vitt et al. 2002), i.e. within the cognitive band of 10 s defined by Newell (1990), whatever the level of aggregation of data and the number of themes involved. OLAP users can focus on the information they are looking for rather than focusing on the process of getting this information, as it used to be with the SQL queries typical of transactional databases. Multidimensional databases nowadays represent a very important aspect of decision-support systems, and are now penetrating the field of GIS in academia (see, for instance, Miller and Han 2001, Bédard et al. 2003, Bédard 2005) but also with commercial products (e.g. JMap SOLAP). Multidimensional databases are very well suited to facilitate data quality analysis in data-rich GIS applications. They are built to specifically query data at the required level of granularity, to provide fast results from complex queries on large volumes of data (not interrupting users train-of-thought), and to allow an intuitive navigation into summarized or detailed interrelated information using different operators (providing interaction with the system). Devillers et al. (2005) presented in detail a model named Quality Information Management Model (QIMM) allowing the management of spatial data quality information within a datacube. Spatial data quality information stored within the QIMM model is afterwards manipulated using Spatial On-Line Analytical Processing (SOLAP; see Rivest et al. 2001, Bédard et al. 2003) to allow the expert to navigate into quality dimensions and to intersect them for any level of detail. The proposed model is based on two dimensions, namely Quality Indicator and Analysed Data, both designed with four levels of granularity (figure 2). Users can explore quality information by navigating within the system at different levels of detail, going for instance along the Analysed Data dimension, from the quality of an entire dataset, down to the quality of a single object instance, and even down to the geometric primitive when available. In each case, the quality may refer to a global Figure 2. Hierarchical dimensions of the quality information management model (the Analysed Data dimension offering alternate paths for detailed analysis).

10 270 R. Devillers et al. indicator or, along the Quality Indicator dimension, go down to a very specific characteristic of quality. For each intersection between these two dimensions, a map displays the region, layers, occurrences, or primitives being analyzed with corresponding visual variables (colour, pattern, weight, etc.) corresponding to quality values. Panning or zooming on the map could trigger an update of the quality indicators displayed in the dashboard. Examples are presented later in this paper. 3.3 Populating the quality database: Combining bottom-up and top-down approaches Once a multidimensional database structure is designed to manage quality information, the next step is to feed this database with existing or derived quality information. Two approaches can be identified: N Bottom-up: This approach is based on existing information describing data quality. It can include metadata distributed by the data producer but can also include any other data quality information that can be assessed. These metadata can be documented at detailed levels (e.g. describing the horizontal spatial accuracy of a specific road segment) and then be aggregated into higherlevel information (e.g. average spatial accuracy of the roads layer of the selected area, i.e. of all roads of this area). As discussed in section 3.1, our approach does not provide one single way to aggregate quality information. The selection of an appropriate aggregation technique depends on the users and uses. The proposed solution lets the quality expert select the aggregation process that best fits their needs. More than one aggregation process can be used in the same datacube if desired. Examples of aggregation techniques are described in more detail in Devillers et al. (2004, 2005) and in most OLAP reference books (e.g. Berson and Smith 1997). N Top-down: This approach involves collecting more global quality information that is not explicitly available, such as an expert s opinion about the average spatial precision of the roads in his region, and in propagating this general level information, when relevant, to detailed levels (e.g. each road of this region could inherit information provided by the experts opinion). For instance, it is typical to see land-surveyors having very good knowledge of a territory and of the quality of the different datasets describing it (e.g. cadastral and topographic data). Using their experience is often the most reliable way to discern that a dataset is relevant or not for various applications in this area. They can also provide insights on the spatial heterogeneity of the quality of certain datasets, identifying sub-regions of higher- and lower-quality in the area covered by the data. The experts can also do this with respect to the period of measurements and other informal criteria (e.g. residential developments subdivided before the end of the 1970s were not measured with Electronic Distance Measurement (EDM) equipment and are not as accurate). Information collected through a top-down process can be a good complement to the bottom-up approach, as many dataset have metadata that are incomplete, too general, or just non-existent. While the two approaches are complementary, they both have advantages and drawbacks. In the bottom-up, metadata can be easier to collect, but finding the most efficient methods to aggregate quality information, to analyse and synthesize hundreds of metadata that vary over space, time, and sources can be difficult. As a

11 single aggregation technique is unlikely to meet every user need, our approach allows the expert to select one or several aggregation techniques that better fit the task at hand. On the other hand, formalizing expert opinions (top-down approach) is not simple either, and the propagation of quality information to lower levels of details has to be done with caution because high-level information can be an implicit aggregation of heterogeneous low-level data. Nevertheless, this remains the most that can be done. With today s knowledge, it seems reasonable to believe that, although none of these approaches can completely fill the quality database, both could be used in most quality information datacubes. The capacity to acquire relevant data becomes a key element when deciding which approach to choose. In addition, in the context of risk analysis for the use of data, one must keep in mind that no information is information, and divergent information is also information. Knowing that no information is available regarding a dataset can lead the user to absorb the remaining uncertainty and proceed to their decision (Bédard 1987). 4. Applying the concepts: Developing the MUM prototype Based on the quality-indicator approach and the QIMM data structure, we developed a prototype to support experts in the assessment of the fitness of certain data for an intended use. The prototype implements, as a proof of concept, different operators which have been described in Devillers et al. (2004), such as displaying quality information using indicators, calculating indicators values according to the spatial extent visualized by the user, allowing users to select indicators relevant to their application, and providing indicators at different levels of details. In the next sections, we describe the architecture of this prototype, the quality indicators that make up the multidimensional data structure and how experts can navigate through the quality information. 4.1 Prototype architecture Quality information analysis tool 271 The prototype was developed using commercial off-the-shelf software driven by a single user interface developed in Visual Basic (fast and easy for prototyping). This user interface adapts and integrates the mapping and database technologies required to suit the needs of the proposed approach (figure 3). The result is a SOLAP application with dashboard capabilities that supports the analysis of spatial data quality. The main technologies include: N Microsoft SQL Server/Analysis Services: This is the OLAP server that provides multidimensional database management functionalities in addition to support OLAP queries in the MDX language; N Microsoft Access: This popular relational database management system is used to store user profiles and multidimensional indicators name and characteristics (indicators metadata); N Proclarity: This OLAP client software provides query and navigation functions (e.g. drill-down and roll-up operators) that allow users to explore the quality data stored within SQL Server; N Intergraph Geomedia Professional: This Geographical Information System (GIS) software provides map-viewing functions such as Zoom In, Zoom Out, Pan, Fit all and other tools allowing the creation of maps representing data quality.

12 272 R. Devillers et al. Figure 3. MUM prototype general architecture. Data quality information used for the experimentation was based on the ISO international standard (ISO/TC ). However, this was only an implementation choice, and the tool presented in this paper can manage and communicate many other types of information about data quality. For increased speed, quality information is stored within the multidimensional database, or datacube, using a full Multidimensional OLAP data structure (MOLAP), in contrast to other possible Relational OLAP structures (ROLAP) mimicking the former (see Berson and Smith 1997, for more details about the different OLAP architectures). After having completed the design of the complete datacube, we experimented with a subset of the QIMM dimensions to make our proof of concept. Within this prototype, we included the entire Quality Indicator dimension and three levels of detail of the Analysed Data dimension (i.e. dataset, data layer, and object feature instance). 4.2 Indicator selection, calculation, and representation The quality indicator approach we adopted is based on three observations: (1) it is impossible in practice to obtain all detailed metadata, and one cannot only provide a unique value for every aspects of the quality; (2) it is too complex to exhaustively consider all factors at once with their detailed spatial and temporal variability; and (3) not all users evaluate quality based on the same criteria. For instance, certain users will be more interested in spatial accuracy, others in completeness, some in temporal data quality, and so on. For this reason, quality indicators can be selected by users according to their needs. Based on the ISO standard, quality indicators were defined and stored hierarchically within a relational database. The ISO quality elements were used for the prototype, but the approach described in this paper allows users to create their own custom-made indicators, since ISO quality elements may not be sufficient to assess the fitness for use. For instance, if a user is interested in how the dataset ontology (i.e. the concepts represented and their definition and characteristics) fits their own user ontology, they could create one or several indicators that would provide some information on this issue, using certain metrics. For instance, some methods allow the similarity between two ontologies to

13 Quality information analysis tool 273 be measured (e.g. see Brodeur et al. 2003, for an example of what geosemantic proximity measurements can be). Users can build the indicators they want and display them in their dashboard by simply applying a drag-and-drop operation from the indicator list to the dashboard creation tool (figure 4). Each indicator definition is stored within the database, including a description of what it represents, the way it is calculated, warnings related to its interpretation, and its importance as defined by the user (expressed in terms of weight), etc. The user can eventually adapt some items further or add more metadata about the indicators. One may select among different graphical representations to illustrate each indicator (e.g. traffic light, smiley, speed meter). Indicator values are based on the spatial extent of the map being displayed to the quality expert. Indeed, if the user zooms in or pans towards a particular region of interest, quality indicators are recalculated for objects located within this area. 4.3 Navigation into spatial data quality information Using the prototype described in the previous section, geospatial data experts can improve their knowledge of data quality through the use of different navigation tools. As demonstrated in the field of OLAP technology, displaying information at different levels of detail within 10 s allows users to analyze the data without interrupting their train of thought. Figure 5 illustrates the benefits of such a system through different questions a user may have regarding data quality and the different tools offered by the system to help in answering those questions Quality indicator representation. Data quality information is communicated through indicators using various representations (e.g. traffic light, smiley, or speed meter). Quality indicator values can be represented using interactive thematic maps displaying quality values on each feature instance or geometric primitive. Using SOLAP operators, it is then possible to drill the data directly on these maps to access another level of detail of the information. Figure 4. Indicators selection tool (left) with the empty dashboard template and indicators description and graphical representation form (right).

14 274 R. Devillers et al. Figure 5. User mind-stream using the MUM system. A global indicator represents the aggregation of all indicators for the displayed area. Each indicator is the aggregation of sub-indicators, down to detailed metadata when available. In our prototype, the quality dashboard can include up to nine indicators, which is consistent with Miller s rule (Miller 1956) that limits human short-term memory to nine chunks of information. The value of each quality indicator varies according to quality. For instance, an indicator using the trafficlight representation can be green, yellow, red, or white, depending on whether the quality exceeds/meets the needs, is close to the needs, does not reach the needs, or is unknown. Several other types of representation are available for each indicator SOLAP navigation along the analysed data dimension. SOLAP fast drilldown and roll-up capabilities are key elements of the prototype. They allow users to navigate from one level of detail to another along the Analysed Data dimension. For instance, this allows users to obtain quality indicator values for the whole dataset, then look at the quality for a certain feature type (e.g. only roads) and move to a more detailed level again to obtain the quality of a single feature instance (e.g. Main Street ). Figure 6 illustrates this example of navigation. The user interface includes cartographic and SOLAP tools in the upper part, the dashboard with its different indicators on the left, and the cartographic interface on the right. These operators fully exploit the advantages of datacubes, allowing the expert to easily take into consideration the spatial variation in data quality, or to discover new trends or patterns into this spatial variability. For instance, modifying min max values of quality categories on the fly could make it possible to capture new spatial variations not considered in the aggregation algorithms (ex. more imprecision along

15 Quality information analysis tool 275 Figure 6. Navigation along the Analysed Data dimension using two successive drill-down operations. a river) or to pinpoint the outliers and specific cases that diverge from the general tendency in their area (e.g. show those lines with positional inaccuracy of 30 m among those fitting within the 5 m inaccuracy). Thus, the expert can see these outliers in spite of a positional accuracy indicator that is smiling. Similar solutions can be used to extract information that may be hidden in aggregated indicators SOLAP navigation along the quality indicator dimension. In addition to the levels of detail within the data, this approach also allows users to explore data quality along a quality indicator hierarchy. The quality indicators in the dashboard can be drilled down and rolled up. Users can then explore quality at aggregated and detailed levels at will, minimizing information overload and offering powerful interaction between the user and the system. For instance, in the example of figure 7, a user looks first at the higher-level indicators. He realizes that General Quality is only average (i.e. yellow) because of the lower Internal Quality. He can then drilldown into the Internal Quality to see its sub-indicators At this second level, he can wonder why the Logical Consistency indicator is only average, and then drill-down on Logical Consistency to obtain more detail. He finally arrives at the last level of detail available in our prototype and sees that the problem comes from the Topological Consistency. He can then decide if this aspect of data quality is important for his application or not and then decide to either absorb the residual uncertainty or reduce it by, for instance, looking for another dataset (see Bédard 1987 or Hunter 1999, for details on the concepts of uncertainty absorption and reduction).

16 276 R. Devillers et al. Figure 7. Navigation along the Quality Indicator dimension using two successive drilldown operations Indicator mapping. Indicator mapping allows users to obtain an insight into the spatial heterogeneity of a quality indicator. As one limitation of the indicator approach is to aggregate heterogeneous data into a single value, this provides a complementary view on quality information. Metadata often document the average quality (e.g. spatial accuracy) for an entire map sheet, but there can be significant spatial variations in data quality at a more detailed level (e.g. feature instance). Consider a dataset covering a large area (e.g. country) that is the result of the integration of several datasets of various qualities that cover smaller adjacent areas (e.g. states). In such a representation, the user may only get a single value for the quality of the whole dataset and may then underestimate or overestimate quality for specific areas. Using the approach presented in this paper, users can explore quality through the indicators displayed in the dashboard. However, when drilling down on specific quality indicators, a user could lose the global picture. To obtain such information, users would have to obtain quality indicator values successively for each feature instance. Quality indicator mapping aims to address this issue. Quality maps can also use different types of classification according to the distribution of values. Five different ways to create the qualitative classes were implemented: equal count, equal range, standard deviation, custom equal count, and custom equal range. Changing the way to create classes can be useful, for instance, when all data of a certain dataset have similar quality levels. Instead of obtaining the same value (e.g. green) for all feature instances, it is then possible to

17 Quality information analysis tool 277 COLOUR FIGURE Figure 8. Mapping of a quality indicator highlighting the spatial heterogeneity of quality information. highlight features with the lowest and the highest qualities in the distribution (figure 8). 5. Conclusion This approach is primarily intended to support quality/audit reporting performed as a one-shot contractual professional activity required by a large client. Consequently, it is not a system to be maintained by an organization as GIS applications typically are. Rather, it is a solution requiring a datacube to be fed with quality indicators defined by the quality expert according to the specific needs of their client. It is a powerful means by which the quality expert will become able to build their opinion in the face of a complex situation, to document their report and to justify their recommendations which will include remarks about the variations of quality in space, features, and so on. As such, the quality datacube may become an instrument like the legal registry of a public officer and serve in Court. MUM is not a software solution but an approach integrating methodological and technological solutions to support legal issues related to spatial data quality. This paper presents a new innovative approach to help data quality experts and expert users of geospatial data to improve their knowledge about spatial data quality in order to better assess the fitness of data for a given use. In addition to being innovative by adapting indicator theory to spatial data quality communication, this approach innovates by its use of both a multidimensional data structure (QIMM) to manage quality information at different levels of detail and a spatial

18 278 R. Devillers et al. OLAP solution in order to support for the first time a rapid and easy exploration of quality indicators at different levels of detail. Thus one can easily and immediately obtain the desired information about the analysed data for a given area at a given epoch. Spatial data quality information is explored along an analysed data dimension and a quality indicator dimension, in addition to being supported by interactive mapping of data quality parameters. Quality information is communicated to users through contextual indicators displayed on a dashboard that is integrated into the mapping interface. The architecture of a prototype was described, as well as its main functionalities that allow users to navigate into diverse quality information at different levels of detail. This prototype was meant as a proof of the applicability of the proposed concepts, the concepts and the general approach being considered as the important results of this research. A validation of the approach was done through demonstrations of the prototype to different types of users (GIS scientists including specialists in data quality issues, consultants in GIS, data producers, governmental agencies, typical GIS users, etc.). Such presentations were performed through the early stages of the project in order to obtain early feedback from potential users and then adapt the project in consequence. Several of these people also directly tested the prototype in order to provide comments on the concept, design, and how the prototype helped them understand data quality. Users expressed an interest in this approach and found it much more efficient than current metadata to increase users knowledge about data quality and help in assessing the fitness of data for certain uses. It was interesting to see that, in addition to the relevance of the prototype for data users, data producers found it very useful to potentially visualize the quality of their data for management purposes. They also realized through this prototype the importance of documenting data quality information in the metadata. A more advanced validation of the approach would help to better quantify the benefit of this approach compared with traditional metadata distribution. It should, however, be done in real-life size (i.e. different datasets, different users, different contexts of use). However, such a test would require considerable human and financial resources because of the technical complexity and was not possible within the scope of this research project. This aspect is currently under development in a project involving three major organizations and will allow the automation of many procedures that were done manually for this project. Different aspects of this research can be expanded to future research, such as improving the model of user needs/profile and formalizing/integrating expert opinions into the QIMM model (i.e. the bottom-up approach). Some studies carried out in the European REVIGIS project for the management of ordinal preferences can also be explored in order to extend MUM numerical indicators (total order) to a partial order that could better express qualitative presentation of data quality. Finally, it is worth mentioning that once quality information is stored in such a structured database with different levels of detail, quality information then becomes easily accessible and can be used to enhance many other aspects of a GIS application. This represents a step towards the creation of quality-aware GIS, which extends the concept of Unwin s (1995) error-sensitive GIS and of Duckham and McCreadie s (2002) error-aware GIS. We refer to a quality-aware GIS as a GIS with the added capabilities to manage, update, explore, assess, and communicate quality information, the term quality encompassing more than error by also addressing issues related to GIS user

19 Quality information analysis tool 279 contexts and use patterns (e.g. user profile and needs assessment). This is then a further step towards safer GIS. Acknowledgements This research is part of the MUM project and has benefited from financial support from the Canadian Network of Centres of Excellence GEOIDE, the IST/FET programme of the European Community (through the REV!GIS project), the Ministère de la Recherche, de la Science et de la Technologie du Québec, the Canada NSERC Industrial Research Chair in Geospatial Databases for Decision-Support, the Centre for Research in Geomatics, and Université Laval. Thanks are due to Mathieu Lachapelle, who contributed to the prototype development, Dr Evan Edinger, for the editing of the manuscript, and to three anonymous reviewers that provided constructive feedback on the paper. References AALDERS, H.J.G.L., 2002, The registration of quality in a GIS. In Spatial Data Quality, W. Shi, P. Fisher and M.F. Goodchild (Eds), pp (London: Taylor & Francis). AALDERS, H.J.G.L. and MORRISON, J., 1998, Spatial data quality for GIS. In Geographic Information Research: Trans-Atlantic Perspectives, M. Craglia and H. Onsrud (Eds), pp (London: Taylor & Francis). AGUMYA, A. and HUNTER, G.J., 1997, Determining fitness for use of geographic information. ITC Journal, 2, pp AGUMYA, A. and HUNTER, G.J., 1999a, Assessing fitness for use of geographic information: What risk are we prepared to accept in our decisions? In Spatial Accuracy Assessment, Land Information Uncertainty in Natural Resources, K. Lowell and A. Jaton (Eds), pp (Chelsea, MI: Ann Arbor Press). AGUMYA, A. and HUNTER, G.J., 1999b, A risk-based approach to assessing the fitness for use of spatial data. URISA Journal, 11, pp BADDELEY, A., 1997, Human Memory: Theory and Practice (Hove, UK: Psychology Press). BEARD, K., 1989, Use error: the neglected error component. In Proceedings of AUTO- CARTO 9, Baltimore, MD, pp BEARD, K., 1997, Representations of data quality. In Geographic Information Research: Bridging the Atlantic, M. Craglia and H. Couclelis (Eds), pp (London: Taylor & Francis). BEARD, K. and BUTTENFIELD, B., 1999, Detecting and evaluating errors by graphical methods. In Geographical Information Systems, P.A. Longley, M.F. Goodchild, D.J. Maguire and D.W. Rhind (Eds), pp (Chichester, UK: Wiley). BEARD, K. and MACKANESS W., 1993, Visual access to data quality in geographic information systems. Cartographica, 30, pp BÉDARD, Y., 1987, Uncertainties in land information systems databases. In Proceedings of Eighth International Symposium on Computer-Assisted Cartography, Baltimore, MD, pp BÉDARD, Y., 2005, Integrating GIS and OLAP: A New Way to Unlock Geospatial Data for Decision-making. Directions on Location Technology and Business Intelligence Conference, 2 4 May, Philadelphia, PA. BÉDARD, Y., DEVILLERS, R., GERVAIS, M. and JEANSOULIN, R., 2004, Towards multidimensional user manuals for geospatial datasets: Legal issues and their considerations in the design of a technological solution. In Proceedings of the Third International Symposium of Spatial Data Quality (ISSDQ), Vol. 2, Bruck an der Leitha, Austria, pp BÉDARD, Y., GOSSELIN, P., RIVEST, S., PROULX, M.-J., NADEAU, M., LEBEL, G. and GAGNON, M.-F., 2003, Integrating GIS components with knowledge discovery

20 280 R. Devillers et al. technology for environmental health decision support. International Journal of Medical Informatics, 70, pp BÉDARD, Y., RIVEST, S. and PROULX, M.-J., 2006, Spatial on-line analytical processing (SOLAP): Concepts, architectures and solutions from a geomatics engineering perspective. In Data Warehouses and OLAP: Concepts, Architectures and Solutions, IDEA Group Publishing, in press. (Hershey: Pennsylvania). BERSON, A. and SMITH, S.J., 1997, Data Warehousing, Data Mining and OLAP (Data Warehousing/Data Management) (New York: McGraw-Hill). BRODEUR, J., BÉDARD, Y., MOULIN, B. and EDWARDS, G., 2003, Revisiting the concept of geospatial data interoperability with the scope of a human communication process. Transactions in GIS, 7, pp BUTTENFIELD, B. and BEARD, K.M., 1994, Graphical and geographical components of data quality. In Visualization in Geographic Information Systems, H.M. Hearnshaw and D.J. Unwin (Eds), pp (Chichester, UK: Wiley). BUTTENFIELD, B.P., 1993, Representing data quality. Cartographica, 30, pp BUTTENFIELD, B.P. and BEARD, K., 1991, Visualizing the quality of spatial information. In Proceedings of AUTO-CARTO 10, pp CHRISMAN, N.R., 1983, The role of quality information in the long term functioning of a geographical information system. In Proceedings of International Symposium on Automated Cartography (Auto Carto 6) (Ottawa, Canada), pp CODD, E.F., 1993, Providing OLAP (On-line Analytical Processing) to User-Analysts: An IT Mandate Report (Sunnyvale, CA: E.F. Codd & Associates). DASSONVILLE, L., VAUGLIN, F., JAKOBSSON, A. and LUZET, C., 2002, Quality management, data quality and users, metadata for geographical information. In Spatial Data Quality, W. Shi, P.F. Fisher and M.F. Goodchild (Eds), pp (London: Taylor & Francis). DE BRUIN, S., BREGT, A. and VAN DE VEN M., 2001, Assessing fitness for use: the expected value of spatial data sets. International Journal of Geographical Information Science, 15, pp DEVILLERS, R., 2004, Conception d un système multidimensionnel d information sur la qualité des données géospatiales. PhD thesis, Sciences Géomatiques, Université Laval, Canada. DEVILLERS, R., BÉDARD, Y. and GERVAIS, M., 2004, Indicateurs de qualité pour réduire les risques de mauvaise utilisation des données géospatiales. Revue Internationale de Géomatique, 14, pp DEVILLERS, R., BÉDARD, Y. and JEANSOULIN, R., 2005, Multidimensional management of geospatial data quality information for its dynamic use within geographical information systems. Photogrammetric Engineering & Remote Sensing (PE&RS), 71, pp DEVILLERS, R., and JEANSOULIN, R. (Eds), 2006, Fundamentals of Spatial Data Quality (London: ISTE). DRECKI, I., 2002, Visualisation of uncertainty in geographic data. In Spatial Data Quality, W. Shi, P.F. Fisher and M.F. Goodchild (Eds), pp (London: Taylor & Francis). DUCKHAM, M. and MCCREADIE, J.E., 2002, Error-aware GIS development. In Spatial Data Quality, W. Shi, P.F. Fisher and M.F. Goodchild (Eds), pp (London: Taylor & Francis). FERNANDEZ, A., 2000, Les nouveaux tableaux de bord des décideurs (Éditions d organisation). FISHER, P.F., 1994, Animation and sound for the visualization of uncertain spatial information. In Visualization in Geographic Information Systems, H.M. Hearnshaw and D.J. Unwin (Eds), pp (Chichester, UK: Wiley). FISHER, P.F., 2003, Multimedia reporting of the results of natural resource surveys. Transactions in GIS, 7, pp

Multidimensional Management of Geospatial Data Quality Information for its Dynamic Use Within GIS

Multidimensional Management of Geospatial Data Quality Information for its Dynamic Use Within GIS Multidimensional Management of Geospatial Data Quality Information for its Dynamic Use Within GIS Rodolphe Devillers, Yvan Bédard, and Robert Jeansoulin Abstract Metadata should help users to assess the

More information

Business Intelligence. A Presentation of the Current Lead Solutions and a Comparative Analysis of the Main Providers

Business Intelligence. A Presentation of the Current Lead Solutions and a Comparative Analysis of the Main Providers 60 Business Intelligence. A Presentation of the Current Lead Solutions and a Comparative Analysis of the Main Providers Business Intelligence. A Presentation of the Current Lead Solutions and a Comparative

More information

Integrating GIS and BI: a Powerful Way to Unlock Geospatial Data for Decision-Making

Integrating GIS and BI: a Powerful Way to Unlock Geospatial Data for Decision-Making Integrating GIS and BI: a Powerful Way to Unlock Geospatial Data for Decision-Making Professor Yvan Bedard, PhD, P.Eng. Centre for Research in Geomatics Laval Univ., Quebec, Canada National Technical University

More information

OLAP and Data Mining. Data Warehousing and End-User Access Tools. Introducing OLAP. Introducing OLAP

OLAP and Data Mining. Data Warehousing and End-User Access Tools. Introducing OLAP. Introducing OLAP Data Warehousing and End-User Access Tools OLAP and Data Mining Accompanying growth in data warehouses is increasing demands for more powerful access tools providing advanced analytical capabilities. Key

More information

RESEARCH ON THE FRAMEWORK OF SPATIO-TEMPORAL DATA WAREHOUSE

RESEARCH ON THE FRAMEWORK OF SPATIO-TEMPORAL DATA WAREHOUSE RESEARCH ON THE FRAMEWORK OF SPATIO-TEMPORAL DATA WAREHOUSE WANG Jizhou, LI Chengming Institute of GIS, Chinese Academy of Surveying and Mapping No.16, Road Beitaiping, District Haidian, Beijing, P.R.China,

More information

GIS Data Quality and Evaluation. Tomislav Sapic GIS Technologist Faculty of Natural Resources Management Lakehead University

GIS Data Quality and Evaluation. Tomislav Sapic GIS Technologist Faculty of Natural Resources Management Lakehead University GIS Data Quality and Evaluation Tomislav Sapic GIS Technologist Faculty of Natural Resources Management Lakehead University The quality of GIS data gains in importance when considering some of the main

More information

Reading Questions. Lo and Yeung, 2007: 2 19. Schuurman, 2004: Chapter 1. 1. What distinguishes data from information? How are data represented?

Reading Questions. Lo and Yeung, 2007: 2 19. Schuurman, 2004: Chapter 1. 1. What distinguishes data from information? How are data represented? Reading Questions Week two Lo and Yeung, 2007: 2 19. Schuurman, 2004: Chapter 1. 1. What distinguishes data from information? How are data represented? 2. What sort of problems are GIS designed to solve?

More information

Developing Business Intelligence and Data Visualization Applications with Web Maps

Developing Business Intelligence and Data Visualization Applications with Web Maps Developing Business Intelligence and Data Visualization Applications with Web Maps Introduction Business Intelligence (BI) means different things to different organizations and users. BI often refers to

More information

DATA WAREHOUSING AND OLAP TECHNOLOGY

DATA WAREHOUSING AND OLAP TECHNOLOGY DATA WAREHOUSING AND OLAP TECHNOLOGY Manya Sethi MCA Final Year Amity University, Uttar Pradesh Under Guidance of Ms. Shruti Nagpal Abstract DATA WAREHOUSING and Online Analytical Processing (OLAP) are

More information

Course 803401 DSS. Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization

Course 803401 DSS. Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization Oman College of Management and Technology Course 803401 DSS Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization CS/MIS Department Information Sharing

More information

DATA QUALITY AND SCALE IN CONTEXT OF EUROPEAN SPATIAL DATA HARMONISATION

DATA QUALITY AND SCALE IN CONTEXT OF EUROPEAN SPATIAL DATA HARMONISATION DATA QUALITY AND SCALE IN CONTEXT OF EUROPEAN SPATIAL DATA HARMONISATION Katalin Tóth, Vanda Nunes de Lima European Commission Joint Research Centre, Ispra, Italy ABSTRACT The proposal for the INSPIRE

More information

A Visualization is Worth a Thousand Tables: How IBM Business Analytics Lets Users See Big Data

A Visualization is Worth a Thousand Tables: How IBM Business Analytics Lets Users See Big Data White Paper A Visualization is Worth a Thousand Tables: How IBM Business Analytics Lets Users See Big Data Contents Executive Summary....2 Introduction....3 Too much data, not enough information....3 Only

More information

BUILDING OLAP TOOLS OVER LARGE DATABASES

BUILDING OLAP TOOLS OVER LARGE DATABASES BUILDING OLAP TOOLS OVER LARGE DATABASES Rui Oliveira, Jorge Bernardino ISEC Instituto Superior de Engenharia de Coimbra, Polytechnic Institute of Coimbra Quinta da Nora, Rua Pedro Nunes, P-3030-199 Coimbra,

More information

GEOG 482/582 : GIS Data Management. Lesson 10: Enterprise GIS Data Management Strategies GEOG 482/582 / My Course / University of Washington

GEOG 482/582 : GIS Data Management. Lesson 10: Enterprise GIS Data Management Strategies GEOG 482/582 / My Course / University of Washington GEOG 482/582 : GIS Data Management Lesson 10: Enterprise GIS Data Management Strategies Overview Learning Objective Questions: 1. What are challenges for multi-user database environments? 2. What is Enterprise

More information

Database Marketing, Business Intelligence and Knowledge Discovery

Database Marketing, Business Intelligence and Knowledge Discovery Database Marketing, Business Intelligence and Knowledge Discovery Note: Using material from Tan / Steinbach / Kumar (2005) Introduction to Data Mining,, Addison Wesley; and Cios / Pedrycz / Swiniarski

More information

Data Warehouse: Introduction

Data Warehouse: Introduction Base and Mining Group of Base and Mining Group of Base and Mining Group of Base and Mining Group of Base and Mining Group of Base and Mining Group of Base and Mining Group of base and data mining group,

More information

Using reporting and data mining techniques to improve knowledge of subscribers; applications to customer profiling and fraud management

Using reporting and data mining techniques to improve knowledge of subscribers; applications to customer profiling and fraud management Using reporting and data mining techniques to improve knowledge of subscribers; applications to customer profiling and fraud management Paper Jean-Louis Amat Abstract One of the main issues of operators

More information

Chapter 5 Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization

Chapter 5 Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization Turban, Aronson, and Liang Decision Support Systems and Intelligent Systems, Seventh Edition Chapter 5 Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization

More information

IBM Cognos 8 Business Intelligence Analysis Discover the factors driving business performance

IBM Cognos 8 Business Intelligence Analysis Discover the factors driving business performance Data Sheet IBM Cognos 8 Business Intelligence Analysis Discover the factors driving business performance Overview Multidimensional analysis is a powerful means of extracting maximum value from your corporate

More information

Interactive Exploration of Multi granularity Spatial and Temporal Datacubes: Providing Computer Assisted Geovisualization Support

Interactive Exploration of Multi granularity Spatial and Temporal Datacubes: Providing Computer Assisted Geovisualization Support Interactive Exploration of Multi granularity Spatial and Temporal Datacubes: Providing Computer Assisted Geovisualization Support Véronique Beaulieu 1 & Yvan Bédard 2 Laval University Centre for Research

More information

Integrating SAP and non-sap data for comprehensive Business Intelligence

Integrating SAP and non-sap data for comprehensive Business Intelligence WHITE PAPER Integrating SAP and non-sap data for comprehensive Business Intelligence www.barc.de/en Business Application Research Center 2 Integrating SAP and non-sap data Authors Timm Grosser Senior Analyst

More information

DATABASE MANAGEMENT SYSTEM

DATABASE MANAGEMENT SYSTEM REVIEW ARTICLE DATABASE MANAGEMENT SYSTEM Sweta Singh Assistant Professor, Faculty of Management Studies, BHU, Varanasi, India E-mail: sweta.v.singh27@gmail.com ABSTRACT Today, more than at any previous

More information

Data Warehouse Snowflake Design and Performance Considerations in Business Analytics

Data Warehouse Snowflake Design and Performance Considerations in Business Analytics Journal of Advances in Information Technology Vol. 6, No. 4, November 2015 Data Warehouse Snowflake Design and Performance Considerations in Business Analytics Jiangping Wang and Janet L. Kourik Walker

More information

CONTINUOUS DATA WAREHOUSE: CONCEPTS, CHALLENGES AND POTENTIALS

CONTINUOUS DATA WAREHOUSE: CONCEPTS, CHALLENGES AND POTENTIALS Geoinformatics 2004 Proc. 12th Int. Conf. on Geoinformatics Geospatial Information Research: Bridging the Pacific and Atlantic University of Gävle, Sweden, 7-9 June 2004 CONTINUOUS DATA WAREHOUSE: CONCEPTS,

More information

OLAP Visualization Operator for Complex Data

OLAP Visualization Operator for Complex Data OLAP Visualization Operator for Complex Data Sabine Loudcher and Omar Boussaid ERIC laboratory, University of Lyon (University Lyon 2) 5 avenue Pierre Mendes-France, 69676 Bron Cedex, France Tel.: +33-4-78772320,

More information

Requirements engineering for a user centric spatial data warehouse

Requirements engineering for a user centric spatial data warehouse Int. J. Open Problems Compt. Math., Vol. 7, No. 3, September 2014 ISSN 1998-6262; Copyright ICSRS Publication, 2014 www.i-csrs.org Requirements engineering for a user centric spatial data warehouse Vinay

More information

Data Warehousing and OLAP Technology for Knowledge Discovery

Data Warehousing and OLAP Technology for Knowledge Discovery 542 Data Warehousing and OLAP Technology for Knowledge Discovery Aparajita Suman Abstract Since time immemorial, libraries have been generating services using the knowledge stored in various repositories

More information

CHAPTER-24 Mining Spatial Databases

CHAPTER-24 Mining Spatial Databases CHAPTER-24 Mining Spatial Databases 24.1 Introduction 24.2 Spatial Data Cube Construction and Spatial OLAP 24.3 Spatial Association Analysis 24.4 Spatial Clustering Methods 24.5 Spatial Classification

More information

DATA MINING TECHNOLOGY. Keywords: data mining, data warehouse, knowledge discovery, OLAP, OLAM.

DATA MINING TECHNOLOGY. Keywords: data mining, data warehouse, knowledge discovery, OLAP, OLAM. DATA MINING TECHNOLOGY Georgiana Marin 1 Abstract In terms of data processing, classical statistical models are restrictive; it requires hypotheses, the knowledge and experience of specialists, equations,

More information

Chapter 5. Warehousing, Data Acquisition, Data. Visualization

Chapter 5. Warehousing, Data Acquisition, Data. Visualization Decision Support Systems and Intelligent Systems, Seventh Edition Chapter 5 Business Intelligence: Data Warehousing, Data Acquisition, Data Mining, Business Analytics, and Visualization 5-1 Learning Objectives

More information

MOC 20467B: Designing Business Intelligence Solutions with Microsoft SQL Server 2012

MOC 20467B: Designing Business Intelligence Solutions with Microsoft SQL Server 2012 MOC 20467B: Designing Business Intelligence Solutions with Microsoft SQL Server 2012 Course Overview This course provides students with the knowledge and skills to design business intelligence solutions

More information

A Capability Model for Business Analytics: Part 2 Assessing Analytic Capabilities

A Capability Model for Business Analytics: Part 2 Assessing Analytic Capabilities A Capability Model for Business Analytics: Part 2 Assessing Analytic Capabilities The first article of this series presented the capability model for business analytics that is illustrated in Figure One.

More information

Healthcare Measurement Analysis Using Data mining Techniques

Healthcare Measurement Analysis Using Data mining Techniques www.ijecs.in International Journal Of Engineering And Computer Science ISSN:2319-7242 Volume 03 Issue 07 July, 2014 Page No. 7058-7064 Healthcare Measurement Analysis Using Data mining Techniques 1 Dr.A.Shaik

More information

Tracking System for GPS Devices and Mining of Spatial Data

Tracking System for GPS Devices and Mining of Spatial Data Tracking System for GPS Devices and Mining of Spatial Data AIDA ALISPAHIC, DZENANA DONKO Department for Computer Science and Informatics Faculty of Electrical Engineering, University of Sarajevo Zmaja

More information

ADVANCED GEOGRAPHIC INFORMATION SYSTEMS Vol. II - Using Ontologies for Geographic Information Intergration Frederico Torres Fonseca

ADVANCED GEOGRAPHIC INFORMATION SYSTEMS Vol. II - Using Ontologies for Geographic Information Intergration Frederico Torres Fonseca USING ONTOLOGIES FOR GEOGRAPHIC INFORMATION INTEGRATION Frederico Torres Fonseca The Pennsylvania State University, USA Keywords: ontologies, GIS, geographic information integration, interoperability Contents

More information

SPATIAL DATA CLASSIFICATION AND DATA MINING

SPATIAL DATA CLASSIFICATION AND DATA MINING , pp.-40-44. Available online at http://www. bioinfo. in/contents. php?id=42 SPATIAL DATA CLASSIFICATION AND DATA MINING RATHI J.B. * AND PATIL A.D. Department of Computer Science & Engineering, Jawaharlal

More information

COURSE SYLLABUS COURSE TITLE:

COURSE SYLLABUS COURSE TITLE: 1 COURSE SYLLABUS COURSE TITLE: FORMAT: CERTIFICATION EXAMS: 55043AC Microsoft End to End Business Intelligence Boot Camp Instructor-led None This course syllabus should be used to determine whether the

More information

How To Choose A Business Intelligence Toolkit

How To Choose A Business Intelligence Toolkit Background Current Reporting Challenges: Difficulty extracting various levels of data from AgLearn Limited ability to translate data into presentable formats Complex reporting requires the technical staff

More information

A Knowledge Management Framework Using Business Intelligence Solutions

A Knowledge Management Framework Using Business Intelligence Solutions www.ijcsi.org 102 A Knowledge Management Framework Using Business Intelligence Solutions Marwa Gadu 1 and Prof. Dr. Nashaat El-Khameesy 2 1 Computer and Information Systems Department, Sadat Academy For

More information

University of Gaziantep, Department of Business Administration

University of Gaziantep, Department of Business Administration University of Gaziantep, Department of Business Administration The extensive use of information technology enables organizations to collect huge amounts of data about almost every aspect of their businesses.

More information

CHAPTER - 5 CONCLUSIONS / IMP. FINDINGS

CHAPTER - 5 CONCLUSIONS / IMP. FINDINGS CHAPTER - 5 CONCLUSIONS / IMP. FINDINGS In today's scenario data warehouse plays a crucial role in order to perform important operations. Different indexing techniques has been used and analyzed using

More information

Business Intelligence and Healthcare

Business Intelligence and Healthcare Business Intelligence and Healthcare SUTHAN SIVAPATHAM SENIOR SHAREPOINT ARCHITECT Agenda Who we are What is BI? Microsoft s BI Stack Case Study (Healthcare) Who we are Point Alliance is an award-winning

More information

Report on the Dagstuhl Seminar Data Quality on the Web

Report on the Dagstuhl Seminar Data Quality on the Web Report on the Dagstuhl Seminar Data Quality on the Web Michael Gertz M. Tamer Özsu Gunter Saake Kai-Uwe Sattler U of California at Davis, U.S.A. U of Waterloo, Canada U of Magdeburg, Germany TU Ilmenau,

More information

ORACLE OLAP. Oracle OLAP is embedded in the Oracle Database kernel and runs in the same database process

ORACLE OLAP. Oracle OLAP is embedded in the Oracle Database kernel and runs in the same database process ORACLE OLAP KEY FEATURES AND BENEFITS FAST ANSWERS TO TOUGH QUESTIONS EASILY KEY FEATURES & BENEFITS World class analytic engine Superior query performance Simple SQL access to advanced analytics Enhanced

More information

City Data Pipeline. A System for Making Open Data Useful for Cities. stefan.bischof@tuwien.ac.at

City Data Pipeline. A System for Making Open Data Useful for Cities. stefan.bischof@tuwien.ac.at City Data Pipeline A System for Making Open Data Useful for Cities Stefan Bischof 1,2, Axel Polleres 1, and Simon Sperl 1 1 Siemens AG Österreich, Siemensstraße 90, 1211 Vienna, Austria {bischof.stefan,axel.polleres,simon.sperl}@siemens.com

More information

Alejandro Vaisman Esteban Zimanyi. Data. Warehouse. Systems. Design and Implementation. ^ Springer

Alejandro Vaisman Esteban Zimanyi. Data. Warehouse. Systems. Design and Implementation. ^ Springer Alejandro Vaisman Esteban Zimanyi Data Warehouse Systems Design and Implementation ^ Springer Contents Part I Fundamental Concepts 1 Introduction 3 1.1 A Historical Overview of Data Warehousing 4 1.2 Spatial

More information

Aspen Collaborative Demand Manager

Aspen Collaborative Demand Manager A world-class enterprise solution for forecasting market demand Aspen Collaborative Demand Manager combines historical and real-time data to generate the most accurate forecasts and manage these forecasts

More information

Automatic Timeline Construction For Computer Forensics Purposes

Automatic Timeline Construction For Computer Forensics Purposes Automatic Timeline Construction For Computer Forensics Purposes Yoan Chabot, Aurélie Bertaux, Christophe Nicolle and Tahar Kechadi CheckSem Team, Laboratoire Le2i, UMR CNRS 6306 Faculté des sciences Mirande,

More information

IBM Cognos Express Essential BI and planning for midsize companies

IBM Cognos Express Essential BI and planning for midsize companies Data Sheet IBM Cognos Express Essential BI and planning for midsize companies Overview IBM Cognos Express is the first and only integrated business intelligence (BI) and planning solution purposebuilt

More information

ORACLE HYPERION PLANNING

ORACLE HYPERION PLANNING ORACLE HYPERION PLANNING ENTERPRISE WIDE PLANNING, BUDGETING, AND FORECASTING KEY FEATURES Hybrid data model facilitates planning, analysis and commentary Flexible workflow capabilities Reliability with

More information

ASSESSMENT OF VISUALIZATION SOFTWARE FOR SUPPORT OF CONSTRUCTION SITE INSPECTION TASKS USING DATA COLLECTED FROM REALITY CAPTURE TECHNOLOGIES

ASSESSMENT OF VISUALIZATION SOFTWARE FOR SUPPORT OF CONSTRUCTION SITE INSPECTION TASKS USING DATA COLLECTED FROM REALITY CAPTURE TECHNOLOGIES ASSESSMENT OF VISUALIZATION SOFTWARE FOR SUPPORT OF CONSTRUCTION SITE INSPECTION TASKS USING DATA COLLECTED FROM REALITY CAPTURE TECHNOLOGIES ABSTRACT Chris Gordon 1, Burcu Akinci 2, Frank Boukamp 3, and

More information

Business Benefits From Microsoft SQL Server Business Intelligence Solutions How Can Business Intelligence Help You? PTR Associates Limited

Business Benefits From Microsoft SQL Server Business Intelligence Solutions How Can Business Intelligence Help You? PTR Associates Limited Business Benefits From Microsoft SQL Server Business Intelligence Solutions How Can Business Intelligence Help You? www.ptr.co.uk Business Benefits From Microsoft SQL Server Business Intelligence (September

More information

CHAPTER 1 INTRODUCTION

CHAPTER 1 INTRODUCTION 1 CHAPTER 1 INTRODUCTION Exploration is a process of discovery. In the database exploration process, an analyst executes a sequence of transformations over a collection of data structures to discover useful

More information

Information Services for Smart Grids

Information Services for Smart Grids Smart Grid and Renewable Energy, 2009, 8 12 Published Online September 2009 (http://www.scirp.org/journal/sgre/). ABSTRACT Interconnected and integrated electrical power systems, by their very dynamic

More information

Framework for Data warehouse architectural components

Framework for Data warehouse architectural components Framework for Data warehouse architectural components Author: Jim Wendt Organization: Evaltech, Inc. Evaltech Research Group, Data Warehousing Practice. Date: 04/08/11 Email: erg@evaltech.com Abstract:

More information

Use of a Web-Based GIS for Real-Time Traffic Information Fusion and Presentation over the Internet

Use of a Web-Based GIS for Real-Time Traffic Information Fusion and Presentation over the Internet Use of a Web-Based GIS for Real-Time Traffic Information Fusion and Presentation over the Internet SUMMARY Dimitris Kotzinos 1, Poulicos Prastacos 2 1 Department of Computer Science, University of Crete

More information

Visualization methods for patent data

Visualization methods for patent data Visualization methods for patent data Treparel 2013 Dr. Anton Heijs (CTO & Founder) Delft, The Netherlands Introduction Treparel can provide advanced visualizations for patent data. This document describes

More information

ORACLE FINANCIAL SERVICES BALANCE SHEET PLANNING

ORACLE FINANCIAL SERVICES BALANCE SHEET PLANNING ORACLE FINANCIAL SERVICES BALANCE SHEET PLANNING KEY FEATURES AND BENEFITS FEATURES Packaged application with prebuilt industry leading practices Net Interest Margin and balance sheet forecasts using cash

More information

Oracle Hyperion Planning

Oracle Hyperion Planning Oracle Hyperion Planning Oracle Hyperion Planning is an agile planning solution that supports enterprise wide planning, budgeting, and forecasting using desktop, mobile and Microsoft Office interfaces.

More information

A Design and implementation of a data warehouse for research administration universities

A Design and implementation of a data warehouse for research administration universities A Design and implementation of a data warehouse for research administration universities André Flory 1, Pierre Soupirot 2, and Anne Tchounikine 3 1 CRI : Centre de Ressources Informatiques INSA de Lyon

More information

Introduction. Chapter 1

Introduction. Chapter 1 Chapter 1 Introduction The area of fault detection and diagnosis is one of the most important aspects in process engineering. This area has received considerable attention from industry and academia because

More information

When to consider OLAP?

When to consider OLAP? When to consider OLAP? Author: Prakash Kewalramani Organization: Evaltech, Inc. Evaltech Research Group, Data Warehousing Practice. Date: 03/10/08 Email: erg@evaltech.com Abstract: Do you need an OLAP

More information

ORACLE BUSINESS INTELLIGENCE, ORACLE DATABASE, AND EXADATA INTEGRATION

ORACLE BUSINESS INTELLIGENCE, ORACLE DATABASE, AND EXADATA INTEGRATION ORACLE BUSINESS INTELLIGENCE, ORACLE DATABASE, AND EXADATA INTEGRATION EXECUTIVE SUMMARY Oracle business intelligence solutions are complete, open, and integrated. Key components of Oracle business intelligence

More information

Tutorials for Project on Building a Business Analytic Model Using Data Mining Tool and Data Warehouse and OLAP Cubes IST 734

Tutorials for Project on Building a Business Analytic Model Using Data Mining Tool and Data Warehouse and OLAP Cubes IST 734 Cleveland State University Tutorials for Project on Building a Business Analytic Model Using Data Mining Tool and Data Warehouse and OLAP Cubes IST 734 SS Chung 14 Build a Data Mining Model using Data

More information

CONCEPTUALIZING BUSINESS INTELLIGENCE ARCHITECTURE MOHAMMAD SHARIAT, Florida A&M University ROSCOE HIGHTOWER, JR., Florida A&M University

CONCEPTUALIZING BUSINESS INTELLIGENCE ARCHITECTURE MOHAMMAD SHARIAT, Florida A&M University ROSCOE HIGHTOWER, JR., Florida A&M University CONCEPTUALIZING BUSINESS INTELLIGENCE ARCHITECTURE MOHAMMAD SHARIAT, Florida A&M University ROSCOE HIGHTOWER, JR., Florida A&M University Given today s business environment, at times a corporate executive

More information

IMPROVING THE QUALITY OF THE DECISION MAKING BY USING BUSINESS INTELLIGENCE SOLUTIONS

IMPROVING THE QUALITY OF THE DECISION MAKING BY USING BUSINESS INTELLIGENCE SOLUTIONS IMPROVING THE QUALITY OF THE DECISION MAKING BY USING BUSINESS INTELLIGENCE SOLUTIONS Maria Dan Ştefan Academy of Economic Studies, Faculty of Accounting and Management Information Systems, Uverturii Street,

More information

TRENDS IN THE DEVELOPMENT OF BUSINESS INTELLIGENCE SYSTEMS

TRENDS IN THE DEVELOPMENT OF BUSINESS INTELLIGENCE SYSTEMS 9 8 TRENDS IN THE DEVELOPMENT OF BUSINESS INTELLIGENCE SYSTEMS Assist. Prof. Latinka Todoranova Econ Lit C 810 Information technology is a highly dynamic field of research. As part of it, business intelligence

More information

www.ducenit.com Self-Service Business Intelligence: The hunt for real insights in hidden knowledge Whitepaper

www.ducenit.com Self-Service Business Intelligence: The hunt for real insights in hidden knowledge Whitepaper Self-Service Business Intelligence: The hunt for real insights in hidden knowledge Whitepaper Shift in BI usage In this fast paced business environment, organizations need to make smarter and faster decisions

More information

Self-Service Business Intelligence

Self-Service Business Intelligence Self-Service Business Intelligence BRIDGE THE GAP VISUALIZE DATA, DISCOVER TRENDS, SHARE FINDINGS Solgenia Analysis provides users throughout your organization with flexible tools to create and share meaningful

More information

Data Warehousing: A Moderated Panel Discussion

Data Warehousing: A Moderated Panel Discussion Data Warehousing: A Moderated Panel Discussion Summary of discussion Moderator: Robin Way, NW Natural, Portland OR Introduction Rationale The Data Warehousing panel represents a new session format for

More information

Information Visualization WS 2013/14 11 Visual Analytics

Information Visualization WS 2013/14 11 Visual Analytics 1 11.1 Definitions and Motivation Lot of research and papers in this emerging field: Visual Analytics: Scope and Challenges of Keim et al. Illuminating the path of Thomas and Cook 2 11.1 Definitions and

More information

Business Intelligence Systems for Analyzing University Students Data

Business Intelligence Systems for Analyzing University Students Data BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 15, No 1 Sofia 2015 Print ISSN: 1311-9702; Online ISSN: 1314-4081 DOI: 10.1515/cait-2015-0009 Business Intelligence Systems

More information

Data Warehouse design

Data Warehouse design Data Warehouse design Design of Enterprise Systems University of Pavia 21/11/2013-1- Data Warehouse design DATA PRESENTATION - 2- BI Reporting Success Factors BI platform success factors include: Performance

More information

SAP BOBJ. Participants will gain the detailed knowledge necessary to design a dashboard that can be used to facilitate the decision making process.

SAP BOBJ. Participants will gain the detailed knowledge necessary to design a dashboard that can be used to facilitate the decision making process. SAP BOBJ 1. BOX310 SAP Business Objects Dashboards 4.0: Core Participants will gain the detailed knowledge necessary to design a dashboard that can be used to facilitate the decision making process. Participants

More information

Principles and Practices of Data Integration

Principles and Practices of Data Integration Data Integration Data integration is the process of combining data of different themes, content, scale or spatial extent, projections, acquisition methods, formats, schema, or even levels of uncertainty,

More information

BestPractices. Dashboard Design: Key Performance Indicators & Metrics Choosing the right data to display. Thomas W. Gonzalez Managing Director

BestPractices. Dashboard Design: Key Performance Indicators & Metrics Choosing the right data to display. Thomas W. Gonzalez Managing Director BestPractices Dashboard Design: Key Performance Indicators & Metrics Choosing the right data to display. Thomas W. Gonzalez Managing Director BrightPoint Consulting, Inc. October 2005 Introduction This

More information

Business Intelligence Solutions for Gaming and Hospitality

Business Intelligence Solutions for Gaming and Hospitality Business Intelligence Solutions for Gaming and Hospitality Prepared by: Mario Perkins Qualex Consulting Services, Inc. Suzanne Fiero SAS Objective Summary 2 Objective Summary The rise in popularity and

More information

Datawarehousing and Business Intelligence

Datawarehousing and Business Intelligence Datawarehousing and Business Intelligence Vannaratana (Bee) Praruksa March 2001 Report for the course component Datawarehousing and OLAP MSc in Information Systems Development Academy of Communication

More information

Journal of Information Technology Management SIGNS OF IT SOLUTIONS FAILURE: REASONS AND A PROPOSED SOLUTION ABSTRACT

Journal of Information Technology Management SIGNS OF IT SOLUTIONS FAILURE: REASONS AND A PROPOSED SOLUTION ABSTRACT Journal of Information Technology Management ISSN #1042-1319 A Publication of the Association of Management SIGNS OF IT SOLUTIONS FAILURE: REASONS AND A PROPOSED SOLUTION MAJED ABUSAFIYA NEW MEXICO TECH

More information

DEVELOPMENT OF A SOLAP PATRIMONY MANAGEMENT APPLICATION SYSTEM: FEZ MEDINA AS A CASE STUDY

DEVELOPMENT OF A SOLAP PATRIMONY MANAGEMENT APPLICATION SYSTEM: FEZ MEDINA AS A CASE STUDY International Journal of Computer Science and Applications, 2008, Vol. 5, No. 3a, pp 57-66 Technomathematics Research Foundation, DEVELOPMENT OF A SOLAP PATRIMONY MANAGEMENT APPLICATION SYSTEM: FEZ MEDINA

More information

Learning Objectives. Definition of OLAP Data cubes OLAP operations MDX OLAP servers

Learning Objectives. Definition of OLAP Data cubes OLAP operations MDX OLAP servers OLAP Learning Objectives Definition of OLAP Data cubes OLAP operations MDX OLAP servers 2 What is OLAP? OLAP has two immediate consequences: online part requires the answers of queries to be fast, the

More information

Adobe Insight, powered by Omniture

Adobe Insight, powered by Omniture Adobe Insight, powered by Omniture Accelerating government intelligence to the speed of thought 1 Challenges that analysts face 2 Analysis tools and functionality 3 Adobe Insight 4 Summary Never before

More information

Portrait Explorer. Discover new customer opportunities through rapid visual insight

Portrait Explorer. Discover new customer opportunities through rapid visual insight Portrait Explorer TM Discover new customer opportunities through rapid visual insight More than BI Effective-decision making for marketing requires more than just BI tools that deliver static reporting,

More information

Improving Decision Making and Managing Knowledge

Improving Decision Making and Managing Knowledge Improving Decision Making and Managing Knowledge Decision Making and Information Systems Information Requirements of Key Decision-Making Groups in a Firm Senior managers, middle managers, operational managers,

More information

Business Intelligence, Analytics & Reporting: Glossary of Terms

Business Intelligence, Analytics & Reporting: Glossary of Terms Business Intelligence, Analytics & Reporting: Glossary of Terms A B C D E F G H I J K L M N O P Q R S T U V W X Y Z Ad-hoc analytics Ad-hoc analytics is the process by which a user can create a new report

More information

Dong-Joo Kang* Dong-Kyun Kang** Balho H. Kim***

Dong-Joo Kang* Dong-Kyun Kang** Balho H. Kim*** Visualization Issues of Mass Data for Efficient HMI Design on Control System in Electric Power Industry Visualization in Computerized Operation & Simulation Tools Dong-Joo Kang* Dong-Kyun Kang** Balho

More information

2010 Oracle Corporation 1

2010 Oracle Corporation 1 1 Introduction to Geospatial Visualization in Oracle Business Intelligence Enterprise Edition, 11g Product Mgmt., Oracle Business Intelligence and Spatial Agenda Introduction When

More information

OLAP. Business Intelligence OLAP definition & application Multidimensional data representation

OLAP. Business Intelligence OLAP definition & application Multidimensional data representation OLAP Business Intelligence OLAP definition & application Multidimensional data representation 1 Business Intelligence Accompanying the growth in data warehousing is an ever-increasing demand by users for

More information

Outline. What is Big data and where they come from? How we deal with Big data?

Outline. What is Big data and where they come from? How we deal with Big data? What is Big Data Outline What is Big data and where they come from? How we deal with Big data? Big Data Everywhere! As a human, we generate a lot of data during our everyday activity. When you buy something,

More information

Digital Cadastral Maps in Land Information Systems

Digital Cadastral Maps in Land Information Systems LIBER QUARTERLY, ISSN 1435-5205 LIBER 1999. All rights reserved K.G. Saur, Munich. Printed in Germany Digital Cadastral Maps in Land Information Systems by PIOTR CICHOCINSKI ABSTRACT This paper presents

More information

Continuous Spatial Data Warehousing

Continuous Spatial Data Warehousing Continuous Spatial Data Warehousing Taher Omran Ahmed Faculty of Science Aljabal Algharby University Azzentan - Libya Taher.ahmed@insa-lyon.fr Abstract Decision support systems are usually based on multidimensional

More information

POLAR IT SERVICES. Business Intelligence Project Methodology

POLAR IT SERVICES. Business Intelligence Project Methodology POLAR IT SERVICES Business Intelligence Project Methodology Table of Contents 1. Overview... 2 2. Visualize... 3 3. Planning and Architecture... 4 3.1 Define Requirements... 4 3.1.1 Define Attributes...

More information

Exploratory Data Analysis for Ecological Modelling and Decision Support

Exploratory Data Analysis for Ecological Modelling and Decision Support Exploratory Data Analysis for Ecological Modelling and Decision Support Gennady Andrienko & Natalia Andrienko Fraunhofer Institute AIS Sankt Augustin Germany http://www.ais.fraunhofer.de/and 5th ECEM conference,

More information

BUSINESS INTELLIGENCE AS SUPPORT TO KNOWLEDGE MANAGEMENT

BUSINESS INTELLIGENCE AS SUPPORT TO KNOWLEDGE MANAGEMENT ISSN 1804-0519 (Print), ISSN 1804-0527 (Online) www.academicpublishingplatforms.com BUSINESS INTELLIGENCE AS SUPPORT TO KNOWLEDGE MANAGEMENT JELICA TRNINIĆ, JOVICA ĐURKOVIĆ, LAZAR RAKOVIĆ Faculty of Economics

More information

Using Business Intelligence techniques to increase the safety of citizens The Tilburg case. Abstract

Using Business Intelligence techniques to increase the safety of citizens The Tilburg case. Abstract Using Business Intelligence techniques to increase the safety of citizens The Tilburg case Sérgio Pascoal 1, Jorge Barandela 2, Filipe Martins 3, Daniel Silva 4, Miguel Santos 5, Isabel Seruca 6 1) Universidade

More information

ProClarity Analytics Family

ProClarity Analytics Family ProClarity Analytics Platform 6 Product Data Sheet Accelerated understanding The ProClarity Analytics family enables organizations to centrally manage, store and deploy best practices and key performance

More information

Portrait Explorer TM - New Customer Opportunities

Portrait Explorer TM - New Customer Opportunities Portrait Explorer TM Discover new customer opportunities through rapid visual insight More than BI Effective-decision making for marketing requires more than just BI tools that deliver static reporting,

More information

Qlik s Associative Model

Qlik s Associative Model White Paper Qlik s Associative Model See the Whole Story that Lives Within Your Data August, 2015 qlik.com Table of Contents Introduction 3 Qlik s associative model 3 Query-based visualization tools only

More information

ORACLE BUSINESS INTELLIGENCE SUITE ENTERPRISE EDITION PLUS

ORACLE BUSINESS INTELLIGENCE SUITE ENTERPRISE EDITION PLUS ORACLE BUSINESS INTELLIGENCE SUITE ENTERPRISE EDITION PLUS PRODUCT FACTS & FEATURES KEY FEATURES Comprehensive, best-of-breed capabilities 100 percent thin client interface Intelligence across multiple

More information

Spatial data quality assessment in GIS

Spatial data quality assessment in GIS Recent Advances in Geodesy and Geomatics Engineering Spatial data quality assessment in GIS DANIELA CRISTIANA DOCAN Surveying and Cadastre Department Technical University of Civil Engineering Bucharest

More information