Linked Statistical Data Analysis
|
|
|
- Mavis Pearson
- 10 years ago
- Views:
Transcription
1 Linked Statistical Data Analysis Sarven Capadisli 1, Sören Auer 2, Reinhard Riedl 3 1 Universität Leipzig, Institut für Informatik, AKSW, Leipzig, Germany, 2 University of Bonn and Fraunhofer IAIS, Bonn, Germany, 3 Bern University of Applied Sciences, Bern, Switzerland 1 [email protected], 2 [email protected], 3 [email protected] Document ID: Abstract. Linked Data principles are increasingly employed to publish high-fidelity, heterogeneous statistical datasets in a distributed way. Currently, there exists no simple way for researchers, journalists and interested people to compare statistical data retrieved from different data stores on the Web. Given that the RDF Data Cube vocabulary is used to describe statistical data, its use makes it possible to discover and identify statistical data artifacts in a uniform way. In this article, the design and implementation of an application and service is presented, which utilizes federated SPARQL queries to gather statistical data from distributed data stores. The R language for statistical computing is employed to perform statistical analyses and visualizations. The Shiny application and server bridges the front-end Web user interface with R on the server-side in order to compare statistical macrodata, and stores analyses results in RDF for future research. As a result, distributed linked statistical data can be more easily explored and analysed. Keywords: Linked Data, SDMX, Statistics, Statistical database, Data integration, Regression analysis, User interface 1 Introduction Statistical data artifacts and the analyses conducted on the data are fundamental to testing scientific theories about our societies. In order for the society to tract and learn from its own vast knowledge about events and things, it needs to be able to gather statistical information from heterogeneous and distributed sources. This is to uncover insights, make predictions, or build smarter systems that the society needs to progress. This brings us to the core of our research challenge; how do we reliably acquire statistical data in a uniform way and conduct well-formed analyses that is accessible to different types of data consumers and users?
2 This article presents an approach towards this challenge with its contributions using statistical linked dataspaces. The work herein offers a Web based user-interface for researchers, journalists, or interested people to compare statistical data from different sources against each other without having any knowledge of the technology underneath or the expertise to develop themselves. The service, which we built, proceeds with running decentralized (federated) structured queries to retrieve data from various endpoints, runs an analysis on the data, and provides the analysis back to the user. For future research, analysis is stored so that it can be searched for and reused. 2 Background As pointed out in Statistical Linked Dataspaces [1], what linked statistics provide, and in fact enable, are queries across datasets: Given that the dimension concepts are interlinked, one can learn from a certain observation's dimension value, and enable the automation of cross-dataset queries. The RDF Data Cube vocabulary [2] is used to describe multi-dimensional statistical data, and SDMX-RDF as one of the statistical information models. It makes it possible to represent significant amounts of heterogeneous statistical data as Linked Data where they can be discovered and identified in a uniform way. The statistical artifacts that are produced, and which use this vocabulary, are invaluable for statisticians, researchers, and developers. Linked SDMX Data [3] provided templates and tooling to transform SDMX-ML data from statistical agencies to RDF/XML, resulting in linked statistical datasets at 270a.info [4] using the RDF Data Cube vocabulary. In addition to semantically uplifting the original data, information pertaining provenance was kept track using the PROV Ontology [5] at transformation time, while incorporating retrieval time provenance data. 3 Analysis platform for Linked Statistical Data The analysis platform is focused on two goals: 1) a Web user interface for researchers to compare macrodata observations and to view plots and analysis results, 2) caching and storage of that analysis for future research and reuse. Here, we describe the platform at stats.270a.info [20]. 3.1 User interface A web application was created to provide users with a simple interface to conduct regression analysis and display of scatter plot(s). The interface presents three drop-down selection areas for the user: an independent variable, a dependent variable, and a time series. Both, the independent and dependent variables are composed of a list of datasets with observations, and time series are composed of reference periods of those observations. Upon selecting and
3 submitting datasets to compare, the interface then presents a scatter plot with the best line of best fit from a list of linear models that is tested. The points in the scatter plot represent locations, in this case, countries, which happen to have a measure value for both variables, as well as the reference period that was selected by the user. Below the scatter-plot, a table of analysis results is presented. Figure [1] is a screenshot of the user interface. Figure 1: stats.270a.info analysis user interface The datasets are compiled by gathering qb:datasets (an RDF Data Cube class for datasets) from each statistical dataspace at 270a.info. Similarly, the reference periods are derived from calendar intervals e.g., YYYY, YYYY-MM-DD, YYYY-QQ. 3.2 Data Requirements Our expectation from the data is that it is modeled using the RDF Data Cube vocabulary and is well-formed. Specifically, it needs to pass some of the integrity constraints as outlined by the vocabulary specification. For our application, some of the essential checks are that: a unique data structure definition (DSD) is used for a dataset, DSD includes measure (value of each observation), Concept dimensions have code lists, Codes from code list. To compare variables from two datasets, there needs to be an agreement on the concepts that are being matched for in respective observations. Here, the primary concern is about reference areas (locations), and making sure that the
4 comparison made for the observations from dataset x (independent variable) and dataset y (dependent variable) are using concepts that are interlinked (using the property skos:exactmatch). Practically, a concept e.g. Switzerland, from at least one of the dataset's code lists should have an arc to the other. It ensures that there is a reliable degree of confidence that the particular concept is interchangeable. Hence, the measure corresponding to the phenomenon being observed, is about the same location in both datasets. Figure [2] shows available outbound interlinks for the datasets at *.270a.info/. Figure 2: Outbound interlinks for 270a.info datasets One additional requirement from the datasets is that the RDF Data Cube component properties (e.g., dimensions, measures) either use sdmx-dimension:refarea, sdmx-dimension:refperiod, sdmx-measure:obsvalue directly or are rdfs:subpropertyofs. Given decentralized mappings of the statistical datasets (published as SDMX-ML), their commonality is expected to be the use, or a reference to SDMX-RDF properties in order to achieve generalized federated queries without having complete knowledge of the structures of the datasets, but rather only the essential bits. In order to proceed with the analysis, we use the selections made by the user: dataset x and dataset y, reference period, and then gather all observations with corresponding reference areas, and measures (observation values). Only the observations whose values for the reference areas with interlinked concepts area are retained in the final result. 3.3 Application
5 Shiny [22], an R package, along with Shiny server [23] is used to build an interactive web application. A Shiny application was built to essentially allow an interaction between the front-end Web application and R. User inputs are set to trigger an event which is sent to the Shiny server and handled by the application written in R. The application assembles a SPARQL query using the input values and then sends them to stats.270a.info/sparql endpoint which dispatches federated queries to two SPARQL endpoints where the datasets are located. The SPARQL query request is handled by the SPARQL client for R. The query results are retrieved and given to R for statistical data analysis. R generates a scatter plot containing the (in)dependent variables, where each point in the chart is a reference area (e.g., country) for that particular reference period selection. Regression analysis is done where correlation, p-value, and the line of best fit is determined after testing several linear models, and shown in the user interface. 3.4 Federated Queries The challenge in federated queries was compromising between what should be processed remotely and sent over the wire versus handling some of that workload by the parent endpoint. Since one of the requirements was to ensure that the concepts are interlinked at either one of the endpoints, each endpoint had to include each observation's reference area as well as its interlinked concept. The result from both endpoints was first joined and then filtered in order to avoid false negatives. That is, either concept x has a skos:exactmatch relationship to concept y, or vice versa, or concept x and concept y are the same. One quick and simple way to minimize the number of results was to filter out exact matches at each endpoint which did not contain the other dataset's domain name. Hence, minimizing the number of join operations which had to be handled by the parent endpoint. The anatomy of the query is shown in detail at Essentially, the SPARQL Endpoint URI and the dataset URI are the only requirements. 3.5 Analysis caching and storing In order to optimize application reactivity for all users, previously user selected options for analysis are cached in the Shiny server session. In addition to a cache that is closest to the user, results from the federated queries as well as the R analysis, which was previously conducted, is stored back into the RDF store with a SPARQL Update. This serves multiple purposes. In the event that the Shiny server is restarted and the cache is no longer available, previously calculated results in the store can be reused, which is still more cost efficient than making new federated queries. Another reason for storing the results back in the RDF store is to offer them over the stats.270a.info SPARQL endpoint for additional discovery and reuse of
6 analysis for researchers. Interesting use cases from this approach emerge immediately. For instance, a researcher or journalist can investigate analysis that meets their criteria. 3.6 URI patterns The design pattern for the analysis URIs which refer to the data and the analysis is aimed to keep the length as minimal as possible, while leaving a trace to encourage self exploration and reuse. As URIs for both independent and dependent variable are based on datasets, and the reference period is codified, their prefixed names are used instead in the analysis URI to keep them short and friendly: /{prefix}:{dataset}/{prefix}:{refperiod} 3.7 Vocabularies Besides the common vocabularies: RDF, RDFS, XSD, OWL, the RDF Data Cube vocabulary is used to describe multi-dimensional statistical data, and SDMX-RDF for the statistical information model. PROV-O is used for provenance coverage. A statistical vocabulary[37] is created to describe analyses. It contains classes for analyses, summaries and each data row that is retrieved. Some of the properties include: graph (e.g., scatter plot), independent and dependent variables, reference period, sample size, p-value, correlation value, correlation method that is used. 4 Conclusions We believe that the presented work here and the prior Linked SDMX Data effort contributed towards strengthening the relationship between Semantic Web / Linked Data and statistical communities. The availability of the analysis results in a separate storage system makes it possible for interested parties to take advantage of it variety ways. For instance, a JSON serialization of an analysis or the cached scatter plot in SVG format, is ideal for webpage widgets. Analysis can also be dynamically used in articles or wiki pages with all references intact. The reuse of Linked analyses artifacts as well as the approach to collect data from different sources can help us build smarter systems. It can be employed in fact-checking scenarios as well as uncovering decision-making processes, where knowledge from different sources is put to their potential use when combined. References 1. Capadisli, S.: Statistical Linked Dataspaces. Master's thesis, National University of Ireland (2012),
7 2. The RDF Data Cube vocabulary, 3. Capadisli, S., Auer, S. Ngonga Ngomo, A.-C., Linked SDMX Data, Semantic Web Journal (2013), a.info, 5. The PROV Ontology, 6. stats.270a.info, 7. Shiny, 8. Shiny server, 9. Stats Vocab, Appendix Minimum Requirements The application has to be an end-user application The stats analysis system at is intended for researchers, journalists or interested people. They interact with the Web UI to check for correlation between variables across different statistical datasets. Should be under diverse ownership or control The original data in SDMX-ML from statistical agencies: WB (World Bank), TI (Transparency International), ECB (European Central Bank), FAO (Food and Agriculture Organization of the United Nations), OECD (Organisation for Economic Co-operation and Development), and IMF (International Monetary Fund), first had to go through ETL process and then made available at their respective linked dataspaces at *.270a.info with their own SPARQL endpoints. Should be heterogeneous The original data from all of the sources are in three formats: propriety XML, CSV, and SDMX-ML. All of the data and metadata was composed of varying degree of quality and consistency. Should contain substantial quantities of real world data All of the dataspaces contain both, data and metadata close to 800 million triples in total; consisting of dataset observations, concept schemes, codelists, interlinks, VoID, LODStats and other metadata. There are 68 million triples of observations across all datasets. Meaning must be represented using Semantic Web technologies The statistical data is modeled using the RDF Data Cube vocabulary, which describes multi-dimensional statistical data using the SDMX-RDF as its statistical information model. The regression analysis data is expressed using a custom vocabulary for information like sample size, p-value, correlation value and method that was used, and adjusted R-squared for the independent and dependent variables. Data must be manipulated/processed in interesting ways to derive useful information The observation data that was retrieved from two dataspaces was
8 used to conduct a regression analysis. The results of the analysis are stored in a separate RDF store and can be queried at a SPARQL endpoint. This information is stored for researchers and journalists so that they can discover information that's of interest to them. Semantic information processing has to play a central role in achieving things that alternative technologies cannot do as well, or at all; Given that the statistical dimension concepts across linked dataspaces are interlinked, one can learn from a certain observation's dimension value, and enable the automation of cross-dataset queries in a uniform way with RDF and SPARQL. Additional Desirable Features The application provides an attractive and functional Web interface (for human users) Users only need to make three selections from a drop-down form: independent variable, dependent variable, and the reference period. A scatter plot with the regression line is shown along with the analysis data. Provenance information on the analysis is also provided for the users with the "Oh yeah?" link. The application should be scalable The application uses a decent size of statistical linked data that is in the LOD cloud. The application can use more data sources provided that the data is modelled using RDF Data Cube vocabulary and passes integrity checks, and is available over a SPARQL endpoint. Rigorous evaluations have taken place that demonstrate the benefits of semantic technologies, or validate the results obtained. Interlinking of concept identifiers from different dataspaces on the Web, and the availability of federated queries. Novelty, in applying semantic technology to a domain or task that have not been considered before The application conducts federated queries to different statistical linked dataspaces, gathers the necessary information and proceeding with regression analysis. Storing the analysis results for future use makes it possible for interested parties to discover statistically interesting bits in a uniform way using Semantic Web technologies. Functionality is different from or goes beyond pure information retrieval New data is always created based on user-selections. The application has clear commercial potential and/or large existing user base The application can be used for teaching, support researchers and journalists to discover previously calculated analysis, displaying in external sites, in natural language searches. There is a use of dynamic data, perhaps in combination with static information Data which was not previously analyzed is always retrieved in real-time from federated SPARQL endpoints. The results should be as accurate as possible Federated query retrieves all required observation data, and the regression analysis is conducted by R.
LinkZoo: A linked data platform for collaborative management of heterogeneous resources
LinkZoo: A linked data platform for collaborative management of heterogeneous resources Marios Meimaris, George Alexiou, George Papastefanatos Institute for the Management of Information Systems, Research
ON DEMAND ACCESS TO BIG DATA. Peter Haase fluid Operations AG
ON DEMAND ACCESS TO BIG DATA THROUGHSEMANTIC TECHNOLOGIES Peter Haase fluid Operations AG fluid Operations (fluidops) Linked Data & SemanticTechnologies Enterprise Cloud Computing Software company founded
LINKED DATA EXPERIENCE AT MACMILLAN Building discovery services for scientific and scholarly content on top of a semantic data model
LINKED DATA EXPERIENCE AT MACMILLAN Building discovery services for scientific and scholarly content on top of a semantic data model 22 October 2014 Tony Hammond Michele Pasin Background About Macmillan
Semantic Interoperability
Ivan Herman Semantic Interoperability Olle Olsson Swedish W3C Office Swedish Institute of Computer Science (SICS) Stockholm Apr 27 2011 (2) Background Stockholm Apr 27, 2011 (2) Trends: from
DISCOVERING RESUME INFORMATION USING LINKED DATA
DISCOVERING RESUME INFORMATION USING LINKED DATA Ujjal Marjit 1, Kumar Sharma 2 and Utpal Biswas 3 1 C.I.R.M, University Kalyani, Kalyani (West Bengal) India [email protected] 2 Department of Computer
Mining the Web of Linked Data with RapidMiner
Mining the Web of Linked Data with RapidMiner Petar Ristoski, Christian Bizer, and Heiko Paulheim University of Mannheim, Germany Data and Web Science Group {petar.ristoski,heiko,chris}@informatik.uni-mannheim.de
LDIF - Linked Data Integration Framework
LDIF - Linked Data Integration Framework Andreas Schultz 1, Andrea Matteini 2, Robert Isele 1, Christian Bizer 1, and Christian Becker 2 1. Web-based Systems Group, Freie Universität Berlin, Germany [email protected],
Publishing Linked Data Requires More than Just Using a Tool
Publishing Linked Data Requires More than Just Using a Tool G. Atemezing 1, F. Gandon 2, G. Kepeklian 3, F. Scharffe 4, R. Troncy 1, B. Vatant 5, S. Villata 2 1 EURECOM, 2 Inria, 3 Atos Origin, 4 LIRMM,
Linked Open Data Infrastructure for Public Sector Information: Example from Serbia
Proceedings of the I-SEMANTICS 2012 Posters & Demonstrations Track, pp. 26-30, 2012. Copyright 2012 for the individual papers by the papers' authors. Copying permitted only for private and academic purposes.
Towards a Sales Assistant using a Product Knowledge Graph
Towards a Sales Assistant using a Product Knowledge Graph Haklae Kim, Jungyeon Yang, and Jeongsoon Lee Samsung Electronics Co., Ltd. Maetan dong 129, Samsung-ro, Yeongtong-gu, Suwon-si, Gyeonggi-do 443-742,
Revealing Trends and Insights in Online Hiring Market Using Linking Open Data Cloud: Active Hiring a Use Case Study
Revealing Trends and Insights in Online Hiring Market Using Linking Open Data Cloud: Active Hiring a Use Case Study Amar-Djalil Mezaour 1, Julien Law-To 1, Robert Isele 3, Thomas Schandl 2, and Gerd Zechmeister
The Development of the Clinical Trial Ontology to standardize dissemination of clinical trial data. Ravi Shankar
The Development of the Clinical Trial Ontology to standardize dissemination of clinical trial data Ravi Shankar Open access to clinical trials data advances open science Broad open access to entire clinical
MUSYOP: Towards a Query Optimization for Heterogeneous Distributed Database System in Energy Data Management
MUSYOP: Towards a Query Optimization for Heterogeneous Distributed Database System in Energy Data Management Zhan Liu, Fabian Cretton, Anne Le Calvé, Nicole Glassey, Alexandre Cotting, Fabrice Chapuis
Open Data Integration Using SPARQL and SPIN
Open Data Integration Using SPARQL and SPIN A Case Study for the Tourism Domain Antonino Lo Bue, Alberto Machi ICAR-CNR Sezione di Palermo, Italy Research funded by Italian PON SmartCities Dicet-InMoto-Orchestra
PICASSO Big Data Expert Group
PICASSO Big Data Expert Group Sören Auer Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS The three Big Data V Variety is often neglected Quelle: Gesellschaft für Informatik Fraunhofer
Visual Analysis of Statistical Data on Maps using Linked Open Data
Visual Analysis of Statistical Data on Maps using Linked Open Data Petar Ristoski and Heiko Paulheim University of Mannheim, Germany Research Group Data and Web Science {petar.ristoski,heiko}@informatik.uni-mannheim.de
Scholars@Duke Data Consumer's Guide. Aggregating and consuming data from Scholars@Duke profiles March, 2015
Scholars@Duke Data Consumer's Guide Aggregating and consuming data from Scholars@Duke profiles March, 2015 Contents Getting Started with Scholars@Duke Data 1 Who is this Guide for? 1 Why consume Scholars@Duke
D5.3.2b Automatic Rigorous Testing Components
ICT Seventh Framework Programme (ICT FP7) Grant Agreement No: 318497 Data Intensive Techniques to Boost the Real Time Performance of Global Agricultural Data Infrastructures D5.3.2b Automatic Rigorous
BUSINESS VALUE OF SEMANTIC TECHNOLOGY
BUSINESS VALUE OF SEMANTIC TECHNOLOGY Preliminary Findings Industry Advisory Council Emerging Technology (ET) SIG Information Sharing & Collaboration Committee July 15, 2005 Mills Davis Managing Director
TopBraid Insight for Life Sciences
TopBraid Insight for Life Sciences In the Life Sciences industries, making critical business decisions depends on having relevant information. However, queries often have to span multiple sources of information.
Semantic Stored Procedures Programming Environment and performance analysis
Semantic Stored Procedures Programming Environment and performance analysis Marjan Efremov 1, Vladimir Zdraveski 2, Petar Ristoski 2, Dimitar Trajanov 2 1 Open Mind Solutions Skopje, bul. Kliment Ohridski
Conference on Data Quality for International Organizations
Committee for the Coordination of Statistical Activities Conference on Data Quality for International Organizations Newport, Wales, United Kingdom, 27-28 April 2006 Session 5: Tools and practices for collecting
Toward a framework for statistical data integration
Toward a framework for statistical data integration Ba-Lam Do, Peb Ruswono Aryan, Tuan-Dat Trinh, Peter Wetz, Elmar Kiesling, and A Min Tjoa TU Wien, Vienna, Austria {ba.do,peb.aryan,tuan.trinh, peter.wetz,elmar.kiesling,a.tjoa}@tuwien.ac.at
technische universiteit eindhoven WIS & Engineering Geert-Jan Houben
WIS & Engineering Geert-Jan Houben Contents Web Information System (WIS) Evolution in Web data WIS Engineering Languages for Web data XML (context only!) RDF XML Querying: XQuery (context only!) RDFS SPARQL
Lift your data hands on session
Lift your data hands on session Duration: 40mn Foreword Publishing data as linked data requires several procedures like converting initial data into RDF, polishing URIs, possibly finding a commonly used
Ontology based Recruitment Process
Ontology based Recruitment Process Malgorzata Mochol Radoslaw Oldakowski Institut für Informatik AG Netzbasierte Informationssysteme Freie Universität Berlin Takustr. 9, 14195 Berlin, Germany [email protected]
Building Semantic Content Management Framework
Building Semantic Content Management Framework Eric Yen Computing Centre, Academia Sinica Outline What is CMS Related Work CMS Evaluation, Selection, and Metrics CMS Applications in Academia Sinica Concluding
Linked Data Interface, Semantics and a T-Box Triple Store for Microsoft SharePoint
Linked Data Interface, Semantics and a T-Box Triple Store for Microsoft SharePoint Christian Fillies 1 and Frauke Weichhardt 1 1 Semtation GmbH, Geschw.-Scholl-Str. 38, 14771 Potsdam, Germany {cfillies,
The ECB Statistical Data Warehouse: improving data accessibility for all users
The ECB Statistical Data Warehouse: improving data accessibility for all users Gérard Salou 1 Introduction The Directorate General Statistics of the European Central Bank (ECB) is responsible for the efficient
daq, an Ontology for Dataset Quality Information
daq, an Ontology for Dataset Quality Information Jeremy Debattista University of Bonn / Fraunhofer IAIS, Germany [email protected] Christoph Lange University of Bonn / Fraunhofer IAIS,
AN INTEGRATION APPROACH FOR THE STATISTICAL INFORMATION SYSTEM OF ISTAT USING SDMX STANDARDS
Distr. GENERAL Working Paper No.2 26 April 2007 ENGLISH ONLY UNITED NATIONS STATISTICAL COMMISSION and ECONOMIC COMMISSION FOR EUROPE CONFERENCE OF EUROPEAN STATISTICIANS EUROPEAN COMMISSION STATISTICAL
RDF Dataset Management Framework for Data.go.th
RDF Dataset Management Framework for Data.go.th Pattama Krataithong 1,2, Marut Buranarach 1, and Thepchai Supnithi 1 1 Language and Semantic Technology Laboratory National Electronics and Computer Technology
LINKED OPEN DRUG DATA FROM THE HEALTH INSURANCE FUND OF MACEDONIA
LINKED OPEN DRUG DATA FROM THE HEALTH INSURANCE FUND OF MACEDONIA Milos Jovanovik, Bojan Najdenov, Dimitar Trajanov Faculty of Computer Science and Engineering, Ss. Cyril and Methodius University Skopje,
The IBM Cognos Platform
The IBM Cognos Platform Deliver complete, consistent, timely information to all your users, with cost-effective scale Highlights Reach all your information reliably and quickly Deliver a complete, consistent
Using Open Source software and Open data to support Clinical Trial Protocol design
Using Open Source software and Open data to support Clinical Trial Protocol design Nikolaos Matskanis, Joseph Roumier, Fabrice Estiévenart {nikolaos.matskanis, joseph.roumier, fabrice.estievenart}@cetic.be
Semantic Modeling with RDF. DBTech ExtWorkshop on Database Modeling and Semantic Modeling Lili Aunimo
DBTech ExtWorkshop on Database Modeling and Semantic Modeling Lili Aunimo Expected Outcomes You will learn: Basic concepts related to ontologies Semantic model Semantic web Basic features of RDF and RDF
Integrating SharePoint Sites within WebSphere Portal
Integrating SharePoint Sites within WebSphere Portal November 2007 Contents Executive Summary 2 Proliferation of SharePoint Sites 2 Silos of Information 2 Security and Compliance 3 Overview: Mainsoft SharePoint
Lightweight Data Integration using the WebComposition Data Grid Service
Lightweight Data Integration using the WebComposition Data Grid Service Ralph Sommermeier 1, Andreas Heil 2, Martin Gaedke 1 1 Chemnitz University of Technology, Faculty of Computer Science, Distributed
Creating Knowledge From Interlinked Data From Big Data to Smart Data
Creating Knowledge From Interlinked Data From Big Data to Smart Data Prof. Dr. Sören Auer 2 5. 0 9. 2 0 1 4 Fraunhofer-Institut für Intelligente 1 The Web evolves into a Web of Data Linked Open Data Facebook
Supporting Change-Aware Semantic Web Services
Supporting Change-Aware Semantic Web Services Annika Hinze Department of Computer Science, University of Waikato, New Zealand [email protected] Abstract. The Semantic Web is not only evolving into
The IBM Cognos Platform for Enterprise Business Intelligence
The IBM Cognos Platform for Enterprise Business Intelligence Highlights Optimize performance with in-memory processing and architecture enhancements Maximize the benefits of deploying business analytics
SQL Server 2012 Business Intelligence Boot Camp
SQL Server 2012 Business Intelligence Boot Camp Length: 5 Days Technology: Microsoft SQL Server 2012 Delivery Method: Instructor-led (classroom) About this Course Data warehousing is a solution organizations
Models and Architecture for Smart Data Management
1 Models and Architecture for Smart Data Management Pierre De Vettor, Michaël Mrissa and Djamal Benslimane Université de Lyon, CNRS LIRIS, UMR5205, F-69622, France E-mail: [email protected]
Implementing Data Models and Reports with Microsoft SQL Server 2012 MOC 10778
Implementing Data Models and Reports with Microsoft SQL Server 2012 MOC 10778 Course Outline Module 1: Introduction to Business Intelligence and Data Modeling This module provides an introduction to Business
Improving the visualisation of statistics: The use of SDMX as input for dynamic charts on the ECB website
Improving the visualisation of statistics: The use of SDMX as input for dynamic charts on the ECB website Xavier Sosnovsky, Gérard Salou European Central Bank Abstract The ECB has introduced a number of
Cloud-Based Self Service Analytics
Cloud-Based Self Service Analytics Andrew G Naish* Chief Technical Officer, Space-Time Research, Melbourne, Australia [email protected] Abstracts Traditionally, the means by which official
LiDDM: A Data Mining System for Linked Data
LiDDM: A Data Mining System for Linked Data Venkata Narasimha Pavan Kappara Indian Institute of Information Technology Allahabad Allahabad, India [email protected] Ryutaro Ichise National Institute of
Service Oriented Architecture
Service Oriented Architecture Charlie Abela Department of Artificial Intelligence [email protected] Last Lecture Web Ontology Language Problems? CSA 3210 Service Oriented Architecture 2 Lecture Outline
Industry 4.0 and Big Data
Industry 4.0 and Big Data Marek Obitko, [email protected] Senior Research Engineer 03/25/2015 PUBLIC PUBLIC - 5058-CO900H 2 Background Joint work with Czech Institute of Informatics, Robotics and
Mastering the SAP Business Information Warehouse. Leveraging the Business Intelligence Capabilities of SAP NetWeaver. 2nd Edition
Brochure More information from http://www.researchandmarkets.com/reports/2246934/ Mastering the SAP Business Information Warehouse. Leveraging the Business Intelligence Capabilities of SAP NetWeaver. 2nd
ORACLE BUSINESS INTELLIGENCE SUITE ENTERPRISE EDITION PLUS
ORACLE BUSINESS INTELLIGENCE SUITE ENTERPRISE EDITION PLUS PRODUCT FACTS & FEATURES KEY FEATURES Comprehensive, best-of-breed capabilities 100 percent thin client interface Intelligence across multiple
De la Business Intelligence aux Big Data. Marie- Aude AUFAURE Head of the Business Intelligence team Ecole Centrale Paris. 22/01/14 Séminaire Big Data
De la Business Intelligence aux Big Data Marie- Aude AUFAURE Head of the Business Intelligence team Ecole Centrale Paris 22/01/14 Séminaire Big Data 1 Agenda EvoluHon of Business Intelligence SemanHc Technologies
ORACLE BUSINESS INTELLIGENCE SUITE ENTERPRISE EDITION PLUS
Oracle Fusion editions of Oracle's Hyperion performance management products are currently available only on Microsoft Windows server platforms. The following is intended to outline our general product
COURSE SYLLABUS COURSE TITLE:
1 COURSE SYLLABUS COURSE TITLE: FORMAT: CERTIFICATION EXAMS: 55043AC Microsoft End to End Business Intelligence Boot Camp Instructor-led None This course syllabus should be used to determine whether the
GetLOD - Linked Open Data and Spatial Data Infrastructures
GetLOD - Linked Open Data and Spatial Data Infrastructures W3C Linked Open Data LOD2014 Roma, 20-21 February 2014 Stefano Pezzi, Massimo Zotti, Giovanni Ciardi, Massimo Fustini Agenda Context Geoportal
How To Build A Cloud Based Intelligence System
Semantic Technology and Cloud Computing Applied to Tactical Intelligence Domain Steve Hamby Chief Technology Officer Orbis Technologies, Inc. [email protected] 678.346.6386 1 Abstract The tactical
OECD.Stat Web Browser User Guide
OECD.Stat Web Browser User Guide May 2013 May 2013 1 p.10 Search by keyword across themes and datasets p.31 View and save combined queries p.11 Customise dimensions: select variables, change table layout;
Self-Service Business Intelligence
Self-Service Business Intelligence BRIDGE THE GAP VISUALIZE DATA, DISCOVER TRENDS, SHARE FINDINGS Solgenia Analysis provides users throughout your organization with flexible tools to create and share meaningful
COLINDA: Modeling, Representing and Using Scientific Events in the Web of Data
COLINDA: Modeling, Representing and Using Scientific Events in the Web of Data Selver Softic 1, Laurens De Vocht 2, Martin Ebner 1, Erik Mannens 2, and Rik Van de Walle 2 1 Graz University of Technology,
Explorer's Guide to the Semantic Web
Explorer's Guide to the Semantic Web THOMAS B. PASSIN 11 MANNING Greenwich (74 w. long.) contents preface xiii acknowledgments xv about this booh xvii The Semantic Web 1 1.1 What is the Semantic Web? 3
A HUMAN RESOURCE ONTOLOGY FOR RECRUITMENT PROCESS
A HUMAN RESOURCE ONTOLOGY FOR RECRUITMENT PROCESS Ionela MANIU Lucian Blaga University Sibiu, Romania Faculty of Sciences [email protected] George MANIU Spiru Haret University Bucharest, Romania Faculty
VMware vcenter Log Insight User's Guide
VMware vcenter Log Insight User's Guide vcenter Log Insight 1.5 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition.
HydroDesktop Overview
HydroDesktop Overview 1. Initial Objectives HydroDesktop (formerly referred to as HIS Desktop) is a new component of the HIS project intended to address the problem of how to obtain, organize and manage
Presente e futuro del Web Semantico
Sistemi di Elaborazione dell informazione II Corso di Laurea Specialistica in Ingegneria Telematica II anno 4 CFU Università Kore Enna A.A. 2009-2010 Alessandro Longheu http://www.diit.unict.it/users/alongheu
Big Data, Fast Data, Complex Data. Jans Aasman Franz Inc
Big Data, Fast Data, Complex Data Jans Aasman Franz Inc Private, founded 1984 AI, Semantic Technology, professional services Now in Oakland Franz Inc Who We Are (1 (2 3) (4 5) (6 7) (8 9) (10 11) (12
JavaScript and jquery for Data Analysis and Visualization
Brochure More information from http://www.researchandmarkets.com/reports/2766360/ JavaScript and jquery for Data Analysis and Visualization Description: Go beyond design concepts build dynamic data visualizations
Building Web-based Infrastructures for Smart Meters
Building Web-based Infrastructures for Smart Meters Andreas Kamilaris 1, Vlad Trifa 2, and Dominique Guinard 2 1 University of Cyprus, Nicosia, Cyprus 2 ETH Zurich and SAP Research, Switzerland Abstract.
TopBraid Life Sciences Insight
TopBraid Life Sciences Insight In the Life Sciences industries, making critical business decisions depends on having relevant information. However, queries often have to span multiple sources of information.
We have big data, but we need big knowledge
We have big data, but we need big knowledge Weaving surveys into the semantic web ASC Big Data Conference September 26 th 2014 So much knowledge, so little time 1 3 takeaways What are linked data and the
Microsoft Services Exceed your business with Microsoft SharePoint Server 2010
Microsoft Services Exceed your business with Microsoft SharePoint Server 2010 Business Intelligence Suite Alexandre Mendeiros, SQL Server Premier Field Engineer January 2012 Agenda Microsoft Business Intelligence
- a Humanities Asset Management System. Georg Vogeler & Martina Semlak
- a Humanities Asset Management System Georg Vogeler & Martina Semlak Infrastructure to store and publish digital data from the humanities (e.g. digital scholarly editions): Technically: FEDORA repository
IAAA Grupo de Sistemas de Información Avanzados
Upgrading maps with Linked Data Lopez Pellicer Pellicer, Francisco J Lacasta, Javier Rentería, Walter, Universidad de Zaragoza Barrera, Jesús Lopez de Larrinzar, Juan Agudo, Jose M GeoSpatiumLab The Linked
Ampersand and the Semantic Web
Ampersand and the Semantic Web The Ampersand Conference 2015 Lloyd Rutledge The Semantic Web Billions and billions of data units Triples (subject-predicate-object) of URI s Your data readily integrated
Data Mining, Predictive Analytics with Microsoft Analysis Services and Excel PowerPivot
www.etidaho.com (208) 327-0768 Data Mining, Predictive Analytics with Microsoft Analysis Services and Excel PowerPivot 3 Days About this Course This course is designed for the end users and analysts that
