Pipeline Pilot Enterprise Server. Flexible Integration of Disparate Data and Applications. Capture and Deployment of Best Practices



Similar documents
Cheminformatics and Pharmacophore Modeling, Together at Last

Agilent s Kalabie Electronic Lab Notebook (ELN) Product Overview ChemAxon UGM 2008 Agilent Software and Informatics Division Mike Burke

ElegantJ BI. White Paper. The Enterprise Option Reporting Tools vs. Business Intelligence

MicroStrategy Course Catalog

BIOINFORMATICS Supporting competencies for the pharma industry

Microsoft Services Exceed your business with Microsoft SharePoint Server 2010

Dr Alexander Henzing

Data Integration Checklist

ORACLE BUSINESS INTELLIGENCE SUITE ENTERPRISE EDITION PLUS

Cisco Data Preparation

ORACLE BUSINESS INTELLIGENCE SUITE ENTERPRISE EDITION PLUS

An Oracle White Paper June Integration Technologies for Primavera Solutions

Delivering the power of the world s most successful genomics platform

Leading Genomics. Diagnostic. Discove. Collab. harma. Shanghai Cambridge, MA Reykjavik

How To Choose A Business Intelligence Toolkit

ORACLE DATA INTEGRATOR ENTERPRISE EDITION

CUSTOMER Presentation of SAP Predictive Analytics

MatchPoint Technical Features Tutorial Colygon AG Version 1.0

SAS Enterprise Data Integration Server - A Complete Solution Designed To Meet the Full Spectrum of Enterprise Data Integration Needs

KnowledgeSTUDIO HIGH-PERFORMANCE PREDICTIVE ANALYTICS USING ADVANCED MODELING TECHNIQUES

BUSINESSOBJECTS DATA INTEGRATOR

Outlines. Business Intelligence. What Is Business Intelligence? Data mining life cycle

Doctor of Philosophy in Computer Science

Vanguard Knowledge Automation System

Three Open Blueprints For Big Data Success

ORACLE DATA INTEGRATOR ENTERPRISE EDITION

Self-Service Business Intelligence

How To Build A Connector On A Website (For A Nonprogrammer)

Viewpoint ediscovery Services

HPC & Visualization. Visualization and High-Performance Computing

Databricks. A Primer

SERVICE ORIENTED ARCHITECTURE

Databricks. A Primer

ORACLE REAL-TIME DECISIONS

Business Benefits From Microsoft SQL Server Business Intelligence Solutions How Can Business Intelligence Help You? PTR Associates Limited

Discover more, discover faster. High performance, flexible NLP-based text mining for life sciences

Pentaho High-Performance Big Data Reference Configurations using Cisco Unified Computing System

ElegantJ BI. White Paper. Considering the Alternatives Business Intelligence Solutions vs. Spreadsheets

IBM Tivoli Service Request Manager

2012 LABVANTAGE Solutions, Inc. All Rights Reserved.

OpenText Output Transformation Server

IBM Customer Experience Suite and Electronic Forms

How to Choose the Best Web Content Management System for Customer Experience Management:

KnowledgeSEEKER Marketing Edition

Scientific versus Business Workflows

SOFTWARE TESTING TRAINING COURSES CONTENTS

Data Mining for Successful Healthcare Organizations

THOMSON REUTERS CORTELLIS FOR INFORMATICS. REUTERS/ Aly Song

The Purview Solution Integration With Splunk

Data Virtualization and ETL. Denodo Technologies Architecture Brief

Integrating Bioinformatics, Medical Sciences and Drug Discovery

2015 Workshops for Professors

BUSINESSOBJECTS DATA INTEGRATOR

Ensuring Web Service Quality for Service-Oriented Architectures. An Oracle White Paper June 2008

KnowledgeSEEKER POWERFUL SEGMENTATION, STRATEGY DESIGN AND VISUALIZATION SOFTWARE

Foundations of Business Intelligence: Databases and Information Management

Dynamic Decision-Making Web Services Using SAS Stored Processes and SAS Business Rules Manager

Pivotal CRM 6.0. Benefit for your organization : a solution that can support your business needs

Application Test Management and Quality Assurance

speed thought Getting the most of CHEMAXON Integration June 2006 of The Power of at the

Evaluation of Load/Stress tools for Web Applications testing

ETPL Extract, Transform, Predict and Load

THE QLIKVIEW PRODUCT FAMILY

CRGroup Whitepaper: Digging through the Data. Reporting Options in Microsoft Dynamics GP

A Business Process Services Portal

Jitterbit Technical Overview : Microsoft Dynamics CRM

Azure Machine Learning, SQL Data Mining and R

MEng, BSc Applied Computer Science

Business Intelligence Solutions for Gaming and Hospitality

SAS IT Resource Management 3.2

IBM Enterprise Content Management Product Strategy

Practical Data Science with Azure Machine Learning, SQL Data Mining, and R

Technical Information Abstract

IBM SPSS Modeler Professional

Actuate for PeopleSoft

An EVIDENCE-ENHANCED HEALTHCARE ECOSYSTEM for Cancer: I/T perspectives

Espial IPTV Middleware. Evo Solution Whitepaper. <Title> Delivering Interactive, Personalized 3-Screen Services

Emerging Technologies Shaping the Future of Data Warehouses & Business Intelligence

Cloud Self Service Mobile Business Intelligence MAKE INFORMED DECISIONS WITH BIG DATA ANALYTICS, CLOUD BI, & SELF SERVICE MOBILITY OPTIONS

Data Mining. SPSS Clementine Clementine Overview. Spring 2010 Instructor: Dr. Masoud Yaghini. Clementine

Total Exploration & Production: Field Monitoring Case Study

Reporting component for templates, reports and documents. Formerly XML Publisher.

Ultimus Adaptive BPM Suite V8

Data Mining Solutions for the Business Environment

Agenda. SAP BusinessObjects 2012 / Slide 2 Private and Confidential

E-Business Suite Oracle SOA Suite Integration Options

Database Marketing, Business Intelligence and Knowledge Discovery

JBOSS ENTERPRISE APPLICATION PLATFORM MIGRATION GUIDELINES

Fluency With Information Technology CSE100/IMT100

Tax Fraud in Increasing

Unified Batch & Stream Processing Platform

CLUSTER ANALYSIS WITH R

ORACLE HYPERION PLANNING

OWB Users, Enter The New ODI World

Transcription:

overview Pipeline Pilot Enterprise Server Pipeline Pilot Enterprise Server (PPES) is a powerful client-server platform that streamlines the integration and analysis of the vast quantities of data flooding research informatics. PPES is designed to meet critical requirements of the informatics professional: an agile development environment, fast and secure deployment, minimal maintenance costs, and application extensibility. PPES and its graphical orchestration environment Pipeline Pilot Pipeline Pilot, help you maximize the utility of your information resources through enterprise-scale data flow control, extensive data mining capabilities, and effective reporting. Powerful Data Processing Pipeline Pilot Pipeline Pilot provides the ability to graphically lay out the business processes for managing and mining all your data in a manner tailored to your needs. Functionality in Pipeline Pilot Pipeline Pilot is wrapped into individual components that can be used to compose data processing networks, known as protocols, using hundreds of configurable components for operations such as data retrieval, manipulation, computational filtering, and display. You can create additional components by using standard technologies such as SOAP/Web Services, Perl, Java, and basic command-line access to control third-party or in-house software that are critical to your informatics infrastructure. Through this process, the programs you integrate become modular components in the system, and can be reused by others without the need for manual intervention or specific expertise. The PPES platform exposes a Web services layer that allows these protocols to be integrated as part of a service-oriented architecture (SOA) environment or workflow framework. Flexible Integration of Disparate Data and Applications PPES allows you to transform disparate data into organized information necessary for better-informed decisions across the organization. Using a number of standard technologies, you can bring data together from in-house databases, files, and instruments, collected from anywhere across the enterprise. You can read chemistry, sequence, text, imaging, and numeric data from all popular formats and analyze data from multiple sources in real-time. For highly efficient on-demand data processing and reporting, you can dynamically create data marts using the PPES ETL features. Capture and Deployment of Best Practices With Pipeline Pilot Pipeline Pilot you automatically capture, annotate and version your procedures, allowing you to document and reproduce the steps used to achieve a particular result. Your resulting protocols can be published and shared with others to facilitate cooperative development and knowledge transfer. To make your automated process available to an even wider community, you can make Pipeline Pilot Pipeline Pilot protocols available via the browser-based Pipeline Pilot Web Port interface, via portal technologies such as SharePoint, as well as many other deployment environments. Customized Reporting and Web Application Development You can create customized reports that summarize your data analysis or mining protocols with powerful tables, charts, images, and text. With complete control over the layout and content, you can easily interpret and P I P E L I N E PILOT

Pipeline Pilot Enterprise Server Platform communicate your results. By displaying multiple tables, charts and images in a single report, you can see different views of your data, including side-by-side comparisons of information from different sources and data that was processed in alternate ways. You can deploy your reports in a wide variety of ways, including HTML, PDF, Word, and PowerPoint. To make your reports more dynamic you can add interactivity into the reports, allowing you to link within and outside reports, have tooltips, etc. You can further extend the interactivity to create web applications that link several protocols together into a functional unit. This greatly extends the versatility of your protocols, allowing end-users to provide input data and determine the way the data are presented, without writing protocols for themselves. Pipeline Pilot Enterprise Server Clients Products Pipeline Pilot Professional Pipeline Pilot Professional provides access to the full power of PPES. You can customize PPES by creating or modifying components, and easily publish and share them with others. By configuring integration or scripting components to control external tools, you can extend the functionality of PPES and even control it from a command-line interface for batch processing of protocols. Pipeline Pilot Lite Pipeline Pilot Lite allows you to build and customize protocols from any of the components available on your server, with the exception of those for integration and scripting (reserved for Pipeline Pilot Professional users). The graphical interface permits users to see the processing logic of a protocol, as well as its progress. Pipeline Pilot Web Port Pipeline Pilot Web Port s browser-based environment allows you to extend the power of PPES to a broad community of users. The same protocols that Pipeline Pilot Professional and Lite users run are accessible to Pipeline Pilot Web Port users. Web Port users can browse available protocols and parameterize them to run their own data to suit their unique needs. A Pipeline Pilot Web Port Client seat also makes it possible to run PPES through many other environments, such as regular browsers, PDF documents, and clients developed with one of the Client side SDKs. Discovery Studio Visualizer Client Discovery Studio (DS) is comprehensive modeling and simulation environment for life science research. With the DS Visualizer Client, DS components can be integrated and customized to form complex workflows, which can be deployed as a single job in a unified environment, offering an easy and consistent way to process data. The results of any experiment can be visualized directly within the DS environment where you can display stunning 3D graphics, generate charts, graphs, heat maps and more. With a range of sophisticated science, coupled with an unrivaled ability to integrate any proprietary or third-party tools, you can develop customized workflows to streamline your modeling and simulation tasks. P I P E L I N E PILOT

Pipeline Pilot Enterprise Server Platform Pipeline Pilot Enterprise Server Component Integration and Reporting Integration Reporting Incorporate existing in-house or third party programs as computational services in Pipeline Pilot using Java, Perl, VBScript, or SOAP with the Integration Collection. Also, retrieve data from Oracle using ODBC, and JDBC technologies for analysis or reporting, and store results directly back in your own corporate database. Create customized reports that summarize your data analysis or mining protocols with the tables, charts, images, and text you specify. With complete control over the layout and content, you can easily interpret and communicate your results. By displaying multiple tables, charts and images in a single report, you can see different views of your data, including side-by-side comparisons of information from different sources and data that was processed in alternate ways. Bioinformatics Sequence Analysis Gene Expression Analyze, annotate, and compare biological sequences in an environment where modular tools can be graphically linked together to create practical bioinformatics routines. From over 100 components for sequence analysis, use familiar algorithms like BLAST and ClustalW. You can extend the system yourself by making new components with Java or Perl, or by extending our integration with EMBOSS. Visualize, analyze, annotate, and report on gene expression experiments, including the individual target genes. Core functionality is based on BioConductor the open source software for the analysis and comprehension of genomic data. By harnessing the graphical protocol building capabilities of Pipeline Pilot Pipeline Pilot, the Gene Expression Collection allows you to construct complex workflows without writing code, while also making it easy for you to couple Gene Expression analysis with other PPESbased processes, such as sequence analysis, text analytics, and reporting. Cheminformatics Chemistry ADME/Tox Analyze, profile, and manage the compounds in your corporate database with chemically intelligent filters and learning. Compare your compounds directly with a vendor s flat files to identify novel molecules with appropriate profiles determined by calculated properties. Enumerate and analyze combinatorial libraries at high speed using reactions or Markush library definitions. Identify common cores and group hitlists by chemical families. Execute chemical analysis steps by seamlessly using PPES and third-party processes together to automate a complete analysis. Calculate predicted absorption, distribution, metabolism, excretion and toxicity (ADMET) properties for collections of molecules such as synthesis candidates, vendor libraries, and screening collections. The collection includes models for human intestinal absorption, aqueous solubility, blood brain barrier penetration, plasma protein binding, cytochrome P450 2D6 inhibition, and hepatotoxicity.

Pipeline Pilot Enterprise Server Platform Pipeline Pilot Enterprise Server Component Data Mining and Data Modeling Data Modeling Carry out powerful data modeling with this comprehensive suite of learning and data modeling tools, statistical filters, and clustering components optimized for large realworld data sets. This collection of components extends PPES s standard capabilities to include statistics and predictive modeling for data mining applications. Access powerful methods, such as fast data clustering, Bayesian learning, principal component analysis, linear regression, and partial least squares regression. Perform structure- activity modeling, compound clustering, and maximal common substructure search. Advanced Data Modeling Access advanced data modeling methods for Recursive Partitioning (RP) and Multiobjective Pareto Optimization. A variety of RP methods are available in the collection including both single tree and forest of tree learners. The methods can learn on single or multiple response variables. The Pareto Optimization components include methods for multi-objective optimization problems to provide solutions whose criteria trade off amongst two or more partially conflicting goals. R Statistics Perform insightful analyses, create informative graphics, and make educated decisions. The R Statistics Collection includes components that implement statistical methods for data manipulation, clustering, learning, classical and exploratory data analysis. The underlying statistical engine is the widely used public domain R statistics package. Apply R statistical analysis and data manipulation methods to PPES data streams. You can then incorporate output results from R directly into your pipeline for further analysis using other components in the PPES framework. A sampling of learning methods includes partial least squares, neural nets, and support vector machines. You can easily extend the collection by adding in your own R scripts into an R Custom Script component. Plate Data Analytics Access methods to read, write, report, visualize, manipulate and perform calculations on plate data. The Plate Data Analytics collection allows each record on the data pipeline to carry an entire plate and its associated wells, and to perform both plate and well level operations. Harnessing the power of the Pipeline Pilot Pipeline Pilot graphical protocol building capabilities, complex data analysis procedures, such as analyzing the results of a screening experiment, can be easily constructed without the need to develop code. Coupled with the integration collection, this plate-based data can easily be inserted into or retrieved from a database. Imaging Imaging Leverage the power of Pipeline Pilot Pipeline Pilot s workflow automation and visual programming capabilities for image processing tasks. This intuitive, enterprise- accessible software platform provides sophisticated tools for enhancing, processing, analyzing, measuring, and reporting image data. The Collection integrates image data with numerical, chemical, graphical and textual data in a unified computing framework. Advanced Imaging Perform image processing and analysis with capabilities that rival even the most sophisticated image processing packages available. The Advanced Imaging Collection includes a number of additional features and is well suited for nearly any imaging application. The Collection includes image, measurement, and metadata readers for P I P E L I N E PILOT DICOM and HCS vendor specific file formats, in addition to all the industry-standard imaging formats. A range of additional tools allow you to alter morphology, process sequences and stacks; and transform, cluster, and segment imaging data.

Pipeline Pilot Enterprise Server Platform Pipeline Pilot Enterprise Server Component Life Science Modeling and Simulations Catalyst CHARMm Access a comprehensive set of tools for pharmacophore modeling and 3D database management. The renowned technologies of Catalyst are packaged into the PPES platform for creating automated, easy-to-use workflows that streamline your pharmacophore modeling and analyses. Create customized workflows to deploy Catalyst s sophisticated algorithms to perform conformer generation, 3D database creation, pharmacophore hypothesis generation, virtual screening, and more. Perform simulations on biological molecules using a powerful set of components based on the well validated CHARMm engine. This collection of components extends the standard capabilities of PPES to include stable and accurate molecular mechanics and molecular dynamics simulation of proteins, nucleic acids, small molecules, and protein-ligand complexes. Materials Modeling Materials Access Materials Studio s premier modeling capabilities within the PPES data pipelining environment. The Materials Component Collection allows access to analytical capabilities such as Reflex and Reflex QPA, key functionality of the QSAR Plus package, and a utility to integrate scripting applications developed within Materials Studio s scripting API into Pipeline Pilot Pipeline Pilot protocols. Text Mining and Analytics Text Analytics ChemMining Use text mining to gain critical insights that complement your experimental results. Through document search, characterization, and analyses, your processing routines can be augmented with new information. Achieve your most challenging text mining objectives by linking together search, extraction, and characterization steps into automated routines. Integrate literature mining with your existing scientific protocols, and run them interactively or automatically every night. Survey the IP landscape and avoid duplicating research efforts by using ChemMining to perform chemically-intelligent text mining of public and internal documents. The ChemMining Collection lets you find chemical names in text documents and convert them to live chemical structures ready for further manipulation. You can explore a few documents at once and create a report showing the structures found. Alternatively, you can process a large set of documents and create a searchable database of structures found. Combine your chemical structure searches with textual searches for key terminology, such as terms related to disease processes or biological molecules, to perform truly context-specific, chemically-intelligent searching.

Pipeline Pilot Enterprise Server Platform 0408