Powerful Management of Financial Big Data

Similar documents
Big Data - Infrastructure Considerations

Implementing a Digital Video Archive Using XenData Software and a Spectra Logic Archive

Testing Big data is one of the biggest

StorReduce Technical White Paper Cloud-based Data Deduplication

How In-Memory Data Grids Can Analyze Fast-Changing Data in Real Time

I D C T E C H N O L O G Y S P O T L I G H T. T i m e t o S c ale Out, Not Scale Up

Cloud Customer Architecture for Web Application Hosting, Version 2.0

Redefining Microsoft SQL Server Data Management. PAS Specification

Implementing a Digital Video Archive Based on XenData Software

Implementing an Automated Digital Video Archive Based on the Video Edition of XenData Software

Driving the New Paradigm of Software Defined Storage Solutions White Paper July 2014

NetApp Big Content Solutions: Agile Infrastructure for Big Data

Lecture 32 Big Data. 1. Big Data problem 2. Why the excitement about big data 3. What is MapReduce 4. What is Hadoop 5. Get started with Hadoop

Implement Hadoop jobs to extract business value from large and varied data sets

Big Data Use Case. How Rackspace is using Private Cloud for Big Data. Bryan Thompson. May 8th, 2013

I N T E R S Y S T E M S W H I T E P A P E R F O R F I N A N C I A L SERVICES EXECUTIVES. Deploying an elastic Data Fabric with caché

Building Storage-as-a-Service Businesses

White. Paper. EMC Isilon: A Scalable Storage Platform for Big Data. April 2014

Web Application Hosting Cloud Architecture

Diagram 1: Islands of storage across a digital broadcast workflow

Designing a Cloud Storage System

FTP-Stream Data Sheet

Well packaged sets of preinstalled, integrated, and optimized software on select hardware in the form of engineered systems and appliances

Using an In-Memory Data Grid for Near Real-Time Data Analysis

Hitachi Cloud Service for Content Archiving. Delivered by Hitachi Data Systems

The Power And Use of FireScope Unify ESB

CTERA Enterprise File Services Platform Architecture for HP Helion Content Depot

IBM Smartcloud Managed Backup

Jitterbit Technical Overview : Microsoft Dynamics CRM

HGST Object Storage for a New Generation of IT

Redefining Microsoft Exchange Data Management

Oracle s Big Data solutions. Roger Wullschleger. <Insert Picture Here>

Accelerating and Simplifying Apache

Cloud storage is strategically inevitable

The Dangers of Consumer Grade File Sharing in a Compliance Driven World

Business Transformation for Application Providers

Enabling Database-as-a-Service (DBaaS) within Enterprises or Cloud Offerings

Cisco Unified Data Center Solutions for MapR: Deliver Automated, High-Performance Hadoop Workloads

CTERA Cloud Storage Platform Architecture

The Design and Implementation of the Zetta Storage Service. October 27, 2009

Subject: Request for Proposals: Enterprise Systems Backup Solution, # FIT Due: May 14, 2013 at 2:00 PM EST

I D C V E N D O R S P O T L I G H T

Embedded inside the database. No need for Hadoop or customcode. True real-time analytics done per transaction and in aggregate. On-the-fly linking IP

A very short Intro to Hadoop

Big Data at Cloud Scale

HadoopTM Analytics DDN

Amazon Relational Database Service (RDS)

P u b l i c a t i o n N u m b e r : W P R e v. A

REAL-TIME STREAMING ANALYTICS DATA IN, ACTION OUT

OPEN MODERN DATA ARCHITECTURE FOR FINANCIAL SERVICES RISK MANAGEMENT

Flash Storage: Trust, But Verify

Apache Hadoop: The Big Data Refinery

Building Out Your Cloud-Ready Solutions. Clark D. Richey, Jr., Principal Technologist, DoD

Hitachi Cloud Services Delivered by Hitachi Data Systems for Telco Markets

Protecting Big Data Data Protection Solutions for the Business Data Lake

A new innovation to protect, share, and distribute healthcare data

OPTIMIZING PRIMARY STORAGE WHITE PAPER FILE ARCHIVING SOLUTIONS FROM QSTAR AND CLOUDIAN

Microsoft Big Data Solutions. Anar Taghiyev P-TSP

SAP HANA Data Center Intelligence - Overview Presentation

MiaRec. Architecture for SIPREC recording

<Insert Picture Here> Integrating your On-Premise Applications with Cloud Applications

C Examcollection.Premium.Exam.34q

3Si Managed Authentication Services Service Description

SAP HANA Cloud Platform Frequently Asked Questions - Business

Data Integration Checklist

Innovative technology for big data analytics

Accelerating Hadoop MapReduce Using an In-Memory Data Grid

Frequently Asked Questions

Analyzing Big Data with Splunk A Cost Effective Storage Architecture and Solution

Get Your Out of Control SAP Database Under Control:

Eliminating the Need for WAN Acceleration Using the Cloud

Big Data? Definition # 1: Big Data Definition Forrester Research

System Requirement Specifications (SRS)

Testing 3Vs (Volume, Variety and Velocity) of Big Data

OBIEE 11g Analytics Using EMC Greenplum Database

IBM DB2 CommonStore for Lotus Domino, Version 8.3

SOLUTIONS PRODUCTS INDUSTRIES RESOURCES SUPPORT ABOUT US ClearCube Technology, Inc. All rights reserved. Contact Support

SwiftStack Filesystem Gateway Architecture

Virtual Operational Data Store (VODS) A Syncordant White Paper

Managing Big Data with Hadoop & Vertica. A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database

SAP HANA Cloud Platform. Technical Overview Uwe Heinz

Copyright 2011 Sentry Data Systems, Inc. All Rights Reserved. No Unauthorized Reproduction.

Converged, Real-time Analytics Enabling Faster Decision Making and New Business Opportunities

q for Gods Whitepaper Series (Edition 7) Common Design Principles for kdb+ Gateways

How To Manage A Cloud System

Effective Data Integration - where to begin. Bryte Systems

I/O Considerations in Big Data Analytics

The Kentik Data Engine

Clarifications: 1) We are asking for a two week extension in order to provide a detailed response to the requirements outlined in the REI.

Ranch Networks for Hosted Data Centers

Client Overview. Engagement Situation. Key Requirements

Cloud Storage and Backup

Workshop on Hadoop with Big Data

Service Overview CloudCare Online Backup

Reference Model for Cloud Applications CONSIDERATIONS FOR SW VENDORS BUILDING A SAAS SOLUTION

Managing the Unmanageable: A Better Way to Manage Storage

Transcription:

Powerful Management of Financial Big Data TickSmith s solutions are the first to apply the processing power, speed, and capacity of cutting-edge Big Data technology to financial data. We combine open source software with our proprietary software and customized hardware to manage Petabytes of data at a low cost and with powerful indexing, mining and delivery capabilities. Our platforms are resilient, scalable, easy to manage and flexible. We have variants for specific uses. TickSmith s platforms, such as TickVault, can be extended to offer client-specific data storage for retention and compliance. They can also be monetized by firms that manage cross-connect networks by up selling retention/management/delivery of data to internal and external users. For trade and quote data, and more. For FIX messages and trading history. For back-office data. Global trade & quote data available as a service. TickVault TickVault is our big data platform for storing, managing and delivering financial trading data such as trade & quote history. It is ideal for exchanges, trading networks and data vendors who need to cost effectively manage and extract value from mountains of financial data. TickVault is a comprehensive repository for both structured and unstructured data and is ideally suited as the central repository of content in any Quant Cloud application that a financial institution might want to provide to its users. Many suppliers of analytical applications, both server and cloud-based, are already partnering with TickSmith to ensure their analytical applications can source their data from TickVault.

TickVault - Flexible Deployment TickVault is provided as a complete platform as a service or managed and installed on site on Client premises. In the latter case, TickSmith provides and configures all the hardware based on agreed sizings and capacity requirements. When accessed as a service, TickVault is accessed from the Montreal-area Tier-3 data center where it is currently hosted. Access to and from is through the internet, the latter being connected through a dedicated 1 Gbps Internet link. VPN, dedicated links or any other secure mean can also be provided. Multiple deployment options are available: As a managed service, offered from TickSmith s infrastructure As a dedicated managed platform hosted at one or multiple Client data centers As a combination service and managed platform TickVault provides flexible delivery of data through its REST API. A web interface already exists and can be used as the basis for web delivery of data. TickVault s powerful architecture ensures rapid response to small and large data requests and queries and is also ideal for data mining.

TickVault - How It Works TickVault is built around a core of Big Data technology, including the Hadoop Ecosystem. As such it is ideally suited to manage, index, analyze and deliver huge amounts of structured and unstructured data. Data loading and storage The first step is to load the various data sets that Client produces and want to manage with our platforms. Early access to sample data helps establish an accurate sizing of the platform and the effort needed to put in place the various loading processes. TickVault supports both structured and unstructured data and loads data from files and, in the case of real-time feeds, snaps dropped to files. As described in the diagram, the data is first stored as is in a big data friendly format similar to the original format in which the data was received. This ensures that we can retrieve and regenerate files that are true to the originals. The platform has built-in compression and data is stored and managed compressed. Compression ratios for binary data are typically 3 to 1. As part of the storage process, 3 copies of the data are created and stored on distinct hardware within the platform to ensure redundancy and leverage the parallel processing that is inherent to TickVault. The platform has built-in self-healing and replication capabilities meaning that any hardware failure simply requires a replacement of the affected hard disk or server and TickVault will regenerate the missing data from the other copies.

Once loaded the data is readily available to be delivered in its original format through TickVault s delivery mechanisms. TickVault can return data as it was received, and ensure that the data returned is the exact same as what was originally processed. In addition to delivering data as-is, TickVault can be configured to create value-add versions of the data such as roll-ups, adjusted data, statistics and more. Data transformation A powerful feature of TickVault is its ability to transform or "process" data to ensure that the data that is delivered to end-users is accessible in formats that make it readily usable for their specific applications. Examples of processing include: Creation of trade and quote roll-ups/bars at specific intervals (1 minute, 5 minutes, 1 hour, etc) Creation of trade and quote data adjusted for splits, ticker changes and dividends Creation of corrected trade and quote data (i.e. correction messages applied) Creation of derived data such as VWAP fields or currency converted trade and quote data Creation of "snap files" with a trade and quote at a specific interval (every 5 minutes) Creation of files combining data from multiple data sets Creation of format specific files, including combination of fields and output formats Extraction of specific instruments or lists of instruments Extraction of complete exchanges for a specific period (1 day or 5 years, for example) Creation of reports such as market or instrument statistics Data mining and searches, with different levels of complexity, can also be configured as processes The configuration of transformation processes is based on requirements communicated by Client and done by TickSmith staff. Once a process is in place it can be invoked to run on the fly. For access to complete universes of data or to transformed content that is requested often, transformation processes can be set-up to completely transform existing data and store the results as a separate data source. This is referred to as the transformed data cache in the diagram. Monitoring All processes, from loading to delivery, are monitored using an enterprise-class monitoring solution. Alerts are forwarded in real-time to our team for immediate action. Entitlement and control Data and user entitlements are managed by config. files on the TickVault platform, LDAP is also supported. Both can securely be extended to integrate with a client s entitlement system or DB.

Data Delivery Core delivery is done through a fully documented and easy to use REST API which has the ability to deliver data as files and as streams. TickSmith has also built a basic Web interface that allows easy access to files. The Web (HTTPS) interface is built on top of the REST API. An FTP mechanism also allows access to files. Secure FTP (SFTP) and SSH File Transfer Protocol (SCP) will be supported shortly. Example web interface delivery of HiTS data About TickSmith TickSmith offers cloud-based and on-site solutions to exchanges, trading networks and financial institutions that must accumulate huge quantities of financial data. Based on disruptive "big data" technology, our platforms and powerful indexing and retrieval tools now make it possible to cost effectively manage and extract value from mountains of financial data. TickSmith is already managing 1 Petabyte+ of data. Our background is in financial technology. We have technical resources that understand financial institutions environments, its specificities and how to efficiently code corresponding Big Data solutions. We make financial big data usable for exchanges, trading networks and financial institutions Contact us at Info@ticksmith.com for more information. TickSmith is a member of