IMS Test Data Management. Virtual IMS User Group 4 February 2014

Similar documents
IMS Application Retirement Think the Unthinkable

data express DATA SHEET OVERVIEW

Data Masking Secure Sensitive Data Improve Application Quality. Becky Albin Chief IT Architect

Data Masking Checklist

Data Masking: A baseline data security measure

IBM InfoSphere Optim Test Data Management solution for Oracle E-Business Suite

Optimizing the Performance of the Oracle BI Applications using Oracle Datawarehousing Features and Oracle DAC

DBKDA 2012 : The Fourth International Conference on Advances in Databases, Knowledge, and Data Applications

Test Data Management Concepts

Integrated Data Management: Discovering what you may not know

What's New in SAS Data Management

Protecting Data in a Spooky world Data Masking Technologies DAMA RMC Oct. 29, 2014

Managing Third Party Databases and Building Your Data Warehouse

IBM InfoSphere Discovery: The Power of Smarter Data Discovery

IBM InfoSphere Optim Test Data Management

CA Repository for z/os r7.2

IBM InfoSphere Solutions for System z

Trusted, Enterprise QlikViewreporting. data Integration and data Quality (It s all about data)

Datasheet FUJITSU Cloud Monitoring Service

Enterprise Data Integration The Foundation for Business Insight

OWB Users, Enter The New ODI World

5 Ways Informatica Cloud Data Integration Extends PowerCenter and Enables Hybrid IT. White Paper

SAP BusinessObjects Information Steward

Test Data Management. Representative data, compliant for test environments

Performance rule violations usually result in increased CPU or I/O, time to fix the mistake, and ultimately, a cost to the business unit.

Datamaker for Skytap. Provide full-sized environments filled with up-to-date test data in minutes

Oracle BI Application: Demonstrating the Functionality & Ease of use. Geoffrey Francis Naailah Gora

IPI*GrammTech Original Price List

In-memory Tables Technology overview and solutions

Test Data Management. Services Catalog

IBM InfoSphere Optim Test Data Management Solution

Data Warehouse Center Administration Guide

IBM InfoSphere Optim Data Masking solution

OBIEE DEVELOPER RESUME

Lalit Mittal NIIT Technologies White Paper

The Top 10 Things DBAs Should Know About Toad for IBM DB2

University Data Warehouse Design Issues: A Case Study

A WHITE PAPER By Silwood Technology Limited

SAP Data Services 4.X. An Enterprise Information management Solution

ETL-EXTRACT, TRANSFORM & LOAD TESTING

Real time information -Philips case

"Data Manufacturing: A Test Data Management Solution"

Data Domain Discovery in Test Data Management

Cost Effective Data Management for Oracle Utilities Applications

Microsoft SQL Server versus IBM DB2 Comparison Document (ver 1) A detailed Technical Comparison between Microsoft SQL Server and IBM DB2

Informatica PowerCenter Data Virtualization Edition

DB2 for z/os Best Practices with Virtual Provisioning

Building Your EDI Modernization Roadmap

Automated Invoice/P2P Processing

Report and Dashboard Template User Guide

Automated Invoice/P2P Processing

Building Effective Test Data Management In Distributed Environment

Data Masking for Adabas. Becky Albin Chief IT Architect

Application retirement: enterprise data management strategies for decommissioning projects

Migrating Lotus Notes Applications to Google Apps

IBM Sterling Control Center

zenterprise The Ideal Platform For Smarter Computing Developing Hybrid Applications For zenterprise

Tableau Metadata Model

Data Integrator: Object Naming Conventions

PowerExchange HotFix Installation Overview

BUSINESSOBJECTS DATA INTEGRATOR

The Informatica Solution for Data Privacy

TECHNOLOGY BRIEF: CA ERWIN SAPHIR OPTION. CA ERwin Saphir Option

What s New with Informatica Data Services & PowerCenter Data Virtualization Edition

Realizing the Benefits of Data Modernization

Best Practices for Implementing Oracle Data Integrator (ODI) July 21, 2011

Master Your Data and Your Business Using Informatica MDM. Ravi Shankar Sr. Director, MDM Product Marketing

Informatica PowerCenter The Foundation of Enterprise Data Integration

IBM Optim. The ROI of an Archiving Project. Michael Mittman Optim Products IBM Software Group IBM Corporation

Oracle Data Integrator 11g: Integration and Administration

File Manager base component

Service Oriented Data Management

Database migration using Wizard, Studio and Commander. Based on migration from Oracle to PostgreSQL (Greenplum)

Using Oracle Data Integrator with Essbase, Planning and the Rest of the Oracle EPM Products

PeopleTools Tables: The Application Repository in the Database

EMA Radar for Workload Automation (WLA): Q2 2012

Business 360 Online - Product concepts and features

BUSINESSOBJECTS DATA INTEGRATOR

Efficient and Real Time Data Integration With Change Data Capture

Enforcive /Cross-Platform Audit

Enforcive / Enterprise Security

End to End Solution to Accelerate Data Warehouse Optimization. Franco Flore Alliance Sales Director - APJ

Beyond the Single View with IBM InfoSphere

Ganzheitliches Datenmanagement

Taming the Elephant with Big Data Management. Deep Dive

HP Service Manager. Software Version: 9.40 For the supported Windows and Linux operating systems. Request Management help topics for printing

SAS Enterprise Data Integration Server - A Complete Solution Designed To Meet the Full Spectrum of Enterprise Data Integration Needs

60 TB of Savings in 4 Days: EMC IT s Informatica Data Archive Proof of Concept

Migration Guide. Informatica PowerExchange (Version 8.1.1)

Using Metadata Manager for System Impact Analysis in Healthcare

Transcription:

IMS Test Data Management Virtual IMS User Group 4 February 2014 John B Boyle Informatica Software

Abstract This session will look at the issues and challenges facing IT Development related to the creation and maintenance of appropriate test data to be used in the testing of new and modified applications. It will then provide an overview and demo of Informatica s Test Data Management Products, with specific reference to their use when IMS data forms all or part of the required Test Data. This will be supplemented by a review of two recent Proof of Concept exercises undertaken for a European Insurance Company and a European Bank, both of which included IMS Data.

Agenda Who am I and what do I do? Who are Informatica? A little history Issues and Challenges Informatica Test Data Management for IMS Case Studies

Who am I and what do I do? 2005-Now Informatica Software 1983-1988 Barclays Bank 1976-1983 Cummins Engines IMS 1.1.4 1.1.5 IMS 1.2 3.1 1988-2003 BMC Software IMS 3.1 8.1 IMS DBA Software Evaluation SysProg Team Leader 2003-2005 Friends Provident IMS 8.1 9.1 Technical Support Consultant Post Sales Support Manager Pre Sales Consultant IMS 8.1 12.1 Product Specialist Mainframe CDC Replication Capacity Planning Application Performance Software Installation Data Control Clerk Computer Operator Programmer IMS DBA Like IMS Not Old, just Older

Who are Informatica and what do they do? 2009 Acquire Applimation 2003 Acquire Striva 2006 Acquire Similarity Data Profiling Data Quality Data Archive Data Masking Data Subsetting 1999 IPO on Nasdaq Mainframe Connectivity Change Data Capture 1993 Company Founded PowerCenter PowerMart ETL Tool

Maximize & Unleash Information Potential Understand it Integrate it Cleanse it Relate it Secure it Act on it Across infrastructure, applications & devices Fast, flexible, free of lock-in and future-proof Deliver agility Reduce risk Unleash potential

Why does Test Data Management matter? Issues and Challenges

My very first system Engine Serial Number Reporting Serial Number Range Engine Details DBITEM40 ITEM Database DARES001 DARES002 Engine Serial Number Listing

My Very First Program DARES002 (Engine Serial Number Listing) We d just started using IBM s Data Dictionary DAR = Darlington Systems ES = Engine Serial Number Project 002 = Second program registered. Written on coding sheets Keyed (to disk) by Punch Room Complied, Linkedited Not quite first time! Desk Checked Ready to test

My first Testing Problem DARES001 not quite finished I need the file it produces to test my program Need to Generate some Test Data Use a Utility IEBGENER (was IEBPTPCH??) //S01 EXEC PGM=IEBGENER //SYSPRINT DD SYSOUT=* //SYSUT1 DD * 1111111111 2222222211 3333333311 4444444411 /* //SYSUT2 DD DISP=(,CATLG,DELETE),SPACE=(TRK,(1)), // DSN=JBOYLE.TEST.ESLDATA, // UNIT=SYSDA, // DCB=(RECFM=FB,LRECL=30,BLKSIZE=3600) //SYSPRINT DD SYSOUT=* //SYSIN DD * GENERATE MAXFLDS=3,MAXLITS=26 RECORD FIELD=(10,1,ZP,1), FIELD=(20,'ENGINE DESCRIPTION X',,7), FIELD=(6,'300177',ZP,27)

Now I have Data Took almost as long as writing the program! EDIT JBOYLE.TEST.ESLDATA Command ===> =COLS> ----+----1----+----2----+----3 ****** ***************************** 000001 ENGINE DESCRIPTION X " 011111CDCCDC4CCECDCDECDD4E0017 11111F55795504523997396507307F ------------------------------------- 000002 ENGINE DESCRIPTION X " 022221CDCCDC4CCECDCDECDD4E0017 22221F55795504523997396507307F ------------------------------------- 000003 ENGINE DESCRIPTION X " 033331CDCCDC4CCECDCDECDD4E0017 33331F55795504523997396507307F ------------------------------------- 000004 ààà ENGINE DESCRIPTION X " 044441CDCCDC4CCECDCDECDD4E0017 44441F55795504523997396507307F

Things my Test Data Didn t Test Empty Input File Invalid Build Date Invalid Serial Number New Page More than 9 Pages Lots of other things

Do you test your IMS Applications? No Wrong Answer Look for a new job You re going bust Yes Right Answer Proceed to next slide

What data do you test them with? Copy of Production Data Wrong Answer Look for a new job You re going to jail! A subset of Production Data which has been de-personalised Right Answer Proceed to next slide

What do we need and Why? A Subset of Production Data for Testing DASD is cheap, but it isn t free Test runs take too long on live sized datasets and use too much system resource Everyone wants their own test environment Data which does not contain Sensitive Information What is Sensitive? PII Personally Identifiable Information PHI Personal Health Information CCI Credit Card Information Most Countries now have Data Privacy Legislation and Penalties! UK Data protection act High profile data losses = bad publicity Offshore testing and Cloud-based Applications

Issues with creating a subset High Volumes of data to process when building a subset of production data Subset needs to be kept current so this is not a one time process Subset must be consistent with respect to Database defined RI Application defined RI (invisible from Database Metadata) Cross Database RI (even less visible) Cross Platform RI Applications span platforms Packaged Applications Understanding a data model containing 80,000 tables?

Options for Creating a Subset In-House developed Programs and Scripts As Part of initial application design? Developed some time later? Maintained by? What about Packaged Applications like SAP, Oracle E-Biz? Enterprise wide consistency? Vendor Supplied Test Data Management Tool Support for All Platforms, Databases, Files used by the Enterprise? License and Maintenance Cost? Ease of use?

Issues with Data Masking Identifying what needs to be masked Masking all data is counter productive Data masking must be consistent Same issues as for consistent Subsetting data is interrelated by Database and Application RI Masked Data must be representative Testing with Random data will not exercise application logic Masked data must be secure Prevent reversal of masking to give original data Masking must be repeatable Need to refresh test data with new data periodically

Options for Data Masking In-House developed Programs and Scripts As Part of initial application design? Developed some time later? Maintained by? What about Packaged Applications like SAP, JD Edwards? Enterprise wide consistency? Vendor Supplied Test Data Management Tool Support for All Platforms, Databases, Files used by the Enterprise? License and Maintenance Cost? Ease of use

Data Masking Solution Basic Requirements

What Vendor?

What Vendor?

Test Data Management - Enterprise 1. Establish what data items needs to be masked 2. Establish how these data items should be masked 3. Use results of 1 and 2 to establish and document Enterprise Data Masking Policy 4. Build Data Masking Rules for the identified data items 5. Establish where these data item are stored 6. Build Data masking processes to mask these data items 7. Establish Inter-Object relationships and dependencies 8. Build subsetting process to limit volume of data written to test set when we mask the data 9. Execute!

Informatica Test Data Management for IMS And other less important things

TDM Overview Browser based Workbench used to Define the entire Test Data Generation Process Subsetting Which records to move (retain) Masking How to de-personalise the data Generates PowerCenter Mappings Contain the logic to move Source to Target Generates PowerCenter Workflows Physical Connection to Source and Target Workflows use PowerExchange to read and write the data PowerExchange provides access to Mainframe Data, Applications/Packages, and Relational Databases on LUW platform

PowerCenter Overview General Purpose Data Integration Tool Moves data from any source to any target Transforms the data on the way through Graphical Design tool with extensive range of predefined Transformations All objects stored in Metadata Repository Metadata Manager provides full lineage analysis

PowerCenter Objects Source Source Metadata Definition Target Target Metadata Definition Transformation Pre-defined processes sorter, joiner, union, router etc. Mapping Links Source via Transformation Logic to Target Session Controls Execution of a mapping specifies physical connection properties to read source and write target Workflow Controls execution of a series of sessions

PowerExchange Overview Provides PowerCenter processes with access to sources & targets where no native client access is available Relational Source - DB2 Allows Import of Metadata from DB2 Catalog Non-Relational Sources - IMS, IDMS, ADABAS, DATACOM, VSAM, SEQ Requires creation of Datamap to provide Relational View

What does TDM provide A profiling tool Find relationships between data sources Find sensitive data based on format and metadata A workbench To define how to select data to be included in a subset operations To define what to mask To define how to mask specific data items An execution framework To manage the execution of the Masking and Subsetting processes A Dashboard To view the status of your various Test Data Management Projects

TDM Profiling Exploits Informatica s Data Explorer product (also part of our Data Quality Solution) Report on unique/non-unique columns Establish potential primary key and foreign key relationships between objects. Suggest groups of related objects (Entities) which could be processed together to provide consistent subset

Data Subsetting Objects Entity Basic building block for subset creation Contains Main Table, Component Table and Filter Criteria Group Additional Tables where all rows are to be selected Template Collection of Entities and Groups Plan Executes the Subsetting Components Specifies the actual source and target

Data Masking Objects Rules Basic building block for masking Define how a particular data element is to be masked Policy Contain one or more rules Plan Executes the Masking Policy Specifies the actual source and target Can execute Subsetting Template and Masking Policy in the same Plan

What does all this mean? Using PowerCenter you can define reusable processing logic which can be applied to different physical sources/targets TDM further extends this re-usability by allowing the creation of re-usable rules, entities, policies which can also be applied to multiple different sources/targets even in different DBMS structures like DB2 and Oracle Higher Re-use Lower Development Costs

Out of the box masking rules Substitution Repeatable or non-repeatable Unique or Non-Unique Dictionaries supplied or use your own Specific Data Types Phone Number/Credit Card Number/email/url/IP Address Random Blurring Ranges Shuffle Advanced Write your own in PowerCenter Implemented as a re-usable Mapplet

Test Data Management 1. Profile your data to establish inter-relationships 2. Build your Masking Rules 3. Build your subsetting criteria 4. Apply masking rules to the appropriate columns 5. Build execution plan 6. Execute plan to read production data and write test data 7. Repeat step 6 as required to build additional test sets and refresh existing test sets

Does it work for IMS Data? YES! Create an Unload File containing production Data Perform steps 1 to 7 on previous slide Execution Plan reads the Production Unload and writes a subsetted and masked Unload File Use this unload file to load your test databases Keep your job and stay out of jail

Why use Unload Files You don t have to! Can also read and write from/to IMS Database BMP, Batch DLI or ODBA Issues with Unkeyed Segments RULES=FIRST/LAST Clean point in time copy Create multiple subsets from same source Repeatable process Minimises impact on Production Systems Better performance / lower overhead Everyone has high performance Unload/Load utilities If you don t talk to IBM/BMC/CA/???

Case Study Swedish Bank Sophisticated Test Data Management Project Extract production data from IMS, DB2 and Oracle Mask it as required Store it in Oracle. Create application specific test sets by extracting stored data from Oracle and writing it back to source DBMS Initial project started before Informatica acquired Applimation, now migrated to TDM Control application built to facilitate extraction of data

Case Study German Insurance Company Multiple Source on different platforms DB2 and IMS on z/os Oracle on Linux SAP on Unix Combination of Custom masking rules for specific data items and standard masking techniques Some new SAP module specific Accelerators created Entities consisting of IMS DB2 and Oracle objects are required

Case Study: Medium Enterprise Insurance Company Overview A recent audit of this insurance company s data privacy and protection processes revealed that existing methods for procuring data for testing purposes and manual methods to mask sensitive information were non-compliant with existing PCI, PHI, and Sarbanes Oxley (SOX) data privacy requirements. In addition, these processes resulted in higher testing and development costs for new and existing IT investments and significantly increased their risk of an unwanted data breach. Solution In response, the company adopted Informatica Test Data Management to streamline acquisition of realistic and purposeful data to avoid copying entire data sets from production systems for testing purposes. Packaged data masking policies and rules compliant with PCI, PHI, and SOX were also applied. Masked data was validated against required policies before using it for testing purposes Results The company realized greater than 50 percent in time savings using Informatica Test Data management vs. previous methods. The number of defects in testing processes were reduced by 30 percent or more. Usage of Informatica Data Subset increased in time savings over 50 percent to capture optimal test cases.

Want more Information? Contact your friendly Informatica Sales Rep Or email me at jboyle@informatica.com Or visit the Informatica Web site http://www.informatica.com/us/solutions/application-information-lifecyclemanagement/test-data-management/

Bibliography Securosis Research: Understanding and Selecting Data Masking Solutions https://securosis.com/assets/library/reports/understandingmasking _FinalMaster_V3.pdf Gartner Magic Quadrant http://www.gartner.com/technology/reprints.do?id=1-1dc8knj&ct=121221&st=sb Enterprise Strategy Group Report on Informatica Data Masking http://www.informatica.com/images/02162_esg-enterprise-datamasking_ar_en-us.pdf Informatica Case Study http://www.informatica.com/images/02494_accelerating-insurancelegacy-modernization_wp_en-us.pdf

Thanks for Listening We cannot hold a torch to light another's path without brightening our own