InDetail. Grid-Tools Test Data Management
|
|
|
- Phebe Morton
- 10 years ago
- Views:
Transcription
1 InDetail Grid-Tools Test Data Management An InDetail Paper by Bloor Research Author : Philip Howard Publish date : March 2011
2 As far as we know, Grid-Tools is the only specialist vendor in this space. In our view, Datamaker is the most extensive and most complete test data management product that is available on the market today. Philip Howard
3 Executive introduction: major issues in test data management Test Data Management (TDM) is about the provisioning of data for non-production environments, especially for test purposes but also for development, training, quality assurance, the creation of demonstration or other dummy data, and so on. There are essentially three ways to derive such data: you can take a copy of your production database and, where appropriate, mask any sensitive data; or you can subset your data and mask it; or you can generate synthetic data, based on an understanding of the underlying data model, which means that no masking is required. Each of these approaches to TDM has both advantages and disadvantages. Copying your database has the advantage of being relatively simple. However, it is expensive in terms of hardware, license and support costs to have multiple copies of the same database. It is not unknown for companies with large development shops to have upwards of twenty different copies of the same database for development and testing purposes. Sub-setting your database is less expensive. However, it suffers from the same problems as all sampling processes in that you can miss outliers. This is particularly important in development environments because outliers may cause the system to break whereas normal results do not, so it is important that outliers are properly tested. Therefore, if you are using sub-setting then you need to ensure that outliers are captured and represented within the process of creating your subset. However, this pre-supposes that those outliers are present in the production database at the point at which the data is sub-setted. Since it is unlikely that all possible outliers are present at any one time this means that a copied or sub-setted database can never be fully representative of what you need to test. A further issue that occurs with both full and partial database copies is that any sensitive data may need to be masked. Of course, there are many applications that address data that is not sensitive or subject to data privacy laws. On the other hand, there are also large numbers of applications where it is necessary to de-identify data to meet compliance requirements or to protect intellectual property. Where that is the case then data will need to be masked. This is not as simple as it may appear at first sight. In the first instance, you need to identity data that needs to be hidden: typically, this is done by looking for relevant patterns of information such as credit card numbers ending xxxx. This can be done manually but it is an onerous process that is better automated. Related to this point is the question of how the masking is to be achieved. This will depend, at least in part, on why you are doing the masking. For example, you could simply hide a credit card number by replacing each digit with an x (xxxx-xxxx-xxxx-xxxx), which will be fine if you are only concerned with data protection. However, if you want to test a payment application then you will need to work with real (pseudo-) numbers in order to test your applications. Similarly, simple shuffling techniques (for example, replacing zip code with 54321) will not work if your application requires a valid zip code. For test data management you will need to mask in such a way that the data remains valid. Further, it may not be simply a question of identifying what data needs to be masked and then hiding it. This is because you need to ensure that data relationships remain intact during the masking process, otherwise application testing may break down. This will, of course, be dependent on the application but in complex environments it can be critical. For example, a patient has a disease, which has a treatment, which has a consulting physician who practices in a particular hospital and uses a designated operating theatre. If you scramble the data so that a patient with flu ends up having open heart surgery then your software may break down simply because your masking routines have not ensured that important relationships remain intact. So, discovery of these relationships may be essential. It should also be borne in mind that masking is never perfect. In healthcare environments, to continue the preceding example, a determined hacker may still be able to identify individuals, precisely because of the need to retain relationships. In addition, and as another example, your largest customer will still be your largest customer even if he, she or it is not immediately identifiable by name. The third alternative is to generate synthetic data. From an a priori perspective this is preferable to using either of the other two approaches because the dataset can be relatively small, assuming that it is representative, thereby keeping costs down and because there is no requirement for masking. Moreover, there is no requirement to access production A Bloor InDetail Paper Bloor Research
4 Executive introduction: major issues in test data management data, which means no impact on operational performance. However, in order to generate representative synthetic data you do need to have a good understanding of the data relationships that are not only embedded with the database schema (or file system) but also those relationships that are implicit within the data but which are not formally detailed within the schema. In other words you need some sort of discovery process; but then you need a comparable capability to do a good job of masking, for the reasons just discussed. You will also want to be able to be able to include errors within your synthetic data generation, as you will wish to test the software in this respect. A further point is that the world does not stand still: trading patterns change over time and you may want to discover such trends that already exist within your data and project those forward to test against patterns of data that may be applicable in two or three years time. This is clearly something that you cannot do using either copy or subset-based approach but which should be possible with synthetic data creation. We should also note that some vendors claim to be able to generate synthetic data based on subsetting or copying the data and then repeatedly masking it. While this can be used to support load testing that is probably the limit of its value and we would not describe this as synthetic data generation in any real sense of that term. Otherwise, claims for synthetic data generation may be based on nothing more than having seed tables. If you are interested in synthetic data generation you will therefore need to be wary of different vendors calling different things by the same name. It is also worth noting that it is a common misconception that tools such as HP s QuickTest Professional generates data: they do not. Finally, another major issue in test data management and, indeed, testing in general, is that of coverage. What you would really like to achieve is testing of every possible code path with every possible combination of data with a minimum of tests. Unfortunately, that is very far from the experience of most testers. Taking a database copy, for example, often supports no more than 30% coverage and often much less. There are mechanisms (which we will discuss later) available to improve this percentage and reduce duplicated testing but this cannot be eliminated because of the very nature of the production data, not least because production data is not representative of all possible data, as previously discussed. The same problem also applies to sub-setted data. Conversely, the aim of synthetic data is to provide a truly representative dataset but without duplication. When combined with appropriate mechanisms it is possible to get as much as 100% functional coverage and 90% code coverage with an absolute minimum of tests. This is much more thorough than typical development environments (where 50% coverage is nearer the norm) and should result in the production of better code in less time and at less cost, because of the reduced number of tests that need to be run. Anecdotal evidence suggests that the use of synthetic data can reduce testing cycles by as much as one third. Fast facts Grid-Tools provides a complete suite of Test Data Management tools, which we will describe in detail in this report. These can be licensed en masse as the single product Datamaker or on a modular basis. As far as we know the company is the only vendor to support all of the methods just discussed for managing test data. Grid-Tools works with data in both flat files and databases and it can also be used to support SOA and user interface test environments. Key findings In the opinion of Bloor Research the following represent the key facts of which prospective users should be aware with respect to Grid- Tools Test Data Management: Grid-Tools supports database copying with masking, sub-setting combined with masking and the generation of synthetic data with full support for referential integrity. Test data, however derived, is stored in a test data warehouse. This allows you to manipulate, filter and, in the case of synthetic data, re-generate the data without having to access the production database. Re-generation, in particular, supports an agile approach to development and testing Bloor Research A Bloor InDetail Paper
5 Executive introduction: major issues in test data management To support both masking and synthetic data creation, as well as archival, Grid-Tools provides data profiling capabilities in order to understand data relationships that exist within the data, regardless of whether these are explicit or implicit. The product can also link to third party tools that can expose relevant data models. The product has advanced coverage support. Synthetic data may include a defined percentage of errors for testing purposes. Two different data masking products are provided. One uses native database connectors for popular databases and generic connectors otherwise. As a result the former is very fast at masking but the latter not so. The former also has extended functionality. We expect (and hope) to see more data sources moving into the native category as time progresses. There is a module designed specifically to support SOA testing. The bottom line Most vendors offering test data management either do so as an appendage to software suites that are primarily focused on application testing, or as an extension to archival and information lifecycle management products. As far as we know, Grid-Tools is the only specialist vendor in this space. It is not surprising, therefore, that it has capabilities that go beyond any of its competitors. Moreover, that is something of an under-statement. It has capabilities that go a long way beyond those of most of its competitors. Indeed, one very wellknown company that you might think of as a major competitor to Grid-Tools (which it is in some instances) is also a partner, prepared to bring Grid-Tools into its customer base when it cannot meet the needs of those customers. That probably says as much as you need to know about Grid-Tools: if even one of its archrivals is prepared to concede that its offering is more extensive and complete than theirs then it truly deserves the tag of market leading. A workflow engine is built into the product. There is a module available to identify trends within your data (how its profile is changing) that can be used in the generation of data to cater for future requirements. A Bloor InDetail Paper Bloor Research
6 Product availability Figure 1: The range of capabilities provided by Datamaker Datamaker is the company s flagship product and it includes a variety of different modules that cover the various technical areas illustrated in Figure 1. Note that although the diagram shows data quality as an aspect of Datamaker the product does not have data cleansing capabilities, though it does include data profiling and when you generate data you can deliberately include errors and duplicates for testing purposes. We will discuss these features in due course. The individual modules are available separately if required. What will not be obvious from this diagram is that there is a module specifically for SOA environments that will simulate relevant services along with data creation. Also noteworthy is the fact that there are two data masking products: Fast Data Masking and Simple Data Masking where the main difference (there are others, which we will discuss) is that the former includes native drivers (for DB2 z/os, IMS, Oracle, SQL Server and Teradata) while the latter is generic. On this topic, Data Archive supports DB2, MySQL, Oracle, SQL Server and Sybase while Datamaker as a whole also supports DB2 UDB, DB2 400, Ingres, PostgreSQL, Sybase ASE (IQ is planned), Informix and InterSystems Caché plus various flat file formats including Excel, VSAM/ISAM, CSV, TXT, SQL and fixed definition files. Synthetic data can be generated for HTML files but these cannot be sub-setted or masked. XML files may be created and masked but not sub-setted and this is also true for HIPAA (40-10, and X12), EDI and SWIFT files. The products run on Windows, Linux, UNIX, IBM i-series and IBM System z. The current version number of the products is Bloor Research A Bloor InDetail Paper
7 Product details As noted previously, there are basically three approaches to test data: copy your database, subset it or generate synthetic data. Both of the first two will often require masking but in order to understand what needs masking you will need to profile your data, and you will similarly need to profile your data in order to support the creation of synthetic data, though for somewhat different reasons: in order to understand the data model underlying your environment. So, we will start with a discussion of Datamaker s profiling capabilities, followed by a consideration of its masking and sub-setting capabilities. If you are using a subset-based approach then you will normally subset first and then mask or do them both concurrently. However, if you simply want to work from a copy of your production database then you will only require masking. For this reason we discuss masking before sub-setting even though that may appear illogical. Finally, we will go on to discuss the various requirements of synthetic data generation as well as other aspects of Grid-Tools offering. Data profiling Stand-alone data profiling tools are widely used for two purposes: to discover the quality of data and where there are errors, as a precursor to data cleansing; and to discover relationships that exist within a dataset regardless of whether these relationships are implicit or explicit. As may be imagined, some of these capabilities are specific to data quality environments while some, including all the relationship discovery features, are also relevant to test data management. For obvious reasons, Grid-Tools focuses on the latter and is not concerned with the former. In practice, the data profiling features provided by Grid-Tools look and feel very much like those of data quality vendors. You can, for example, discover the shape of the data. That is, identify maximum and minimum values, calculate standard deviations, and so on. While we will discuss other uses for data shapes later in this paper for the time being it is important to note that this is where data profiling tools usually stop. Conversely, with Datamaker you can compare columns in a database to see if they have the same or very similar shapes, which might suggest a relationship. Data quality tools are typically limited to looking for things like primary/foreign key relationships and, sometimes, the ability to identify potentially redundant columns (all of which Datamaker can do also). Datamaker also includes filters that allow you to automatically recognise column types such as dates, credit card numbers, social security numbers, s, addresses, columns that contain spaces, columns that contain special characters and so on. Where synthetic data is going to be generated, Datamaker s profiling capabilities are designed to supplement any information about the structure of your data that you may have. This may depend on what facilities you already have in place. For example, if you use a data modeling tool such ERWin, ER/Studio or PowerDesigner, then you can import relevant entity-relationship diagrams from those environments. Otherwise you will need to reverse engineer your existing database schema using the facilities provided by Datamaker. This explicit information can then be supplemented with any implicit details discovered during the profiling phase. In some instances, for example in SAP or Oracle (including Siebel, JD Edwards and PeopleSoft) environments, it may make sense to use specialist third-party tools that can generate relevant data models and then import these into the Datamaker environment. Data masking We have already noted that there are a number of masking techniques that are not suitable for use in a test data environment. Datamaker supports these for those who simply have a data privacy concern and want to simply license a data masking product and nothing else. However, that is not the focus of this report and we will not be discussing these techniques, which include blanking, encryption and simple shuffling, amongst others. As we have noted there are two data masking products: Fast Data Masking and Simple Data Masking. In addition to the distinction between native and generic data drivers (which means that fast Data Masking is around 80 times faster, typically masking over 100 million rows per hour) there are also some functional differences. This also applies when masking flat files as opposed to database tables. Both data masking products come with multiple seed tables, with internationalised versions of these tables where appropriate (such as names) and in both cases you can also add your own seed tables and there is multicolumn capability so that, for example, state and zip code will match. Both products also A Bloor InDetail Paper Bloor Research
8 Product details include cross-reference management so that you can, say, retain the same transformations across runs or databases. Similarly, both products provide auditing, allow you to define your own functions and support flat file masking, though in the case of Simple Data Masking this only supports delimited files whereas Fast Data Masking supports all file types. Again, both products can update data directly within the database but Fast Data Masking also offers the option of extracting data into a staging area, passing it through a masked view, which is a technique that is also known as dynamic data masking, and then building shadow tables. Finally, both products support updating of primary keys, though in the case of Simple Data Masking these will need to be disabled whereas they can be rebuilt automatically using Fast Data Masking. Other features built into Fast Data Masking that are not in the Simple product include data profiling (as discussed previously), version control and difference management, common column discovery to ensure that the same mask is applied to matched columns, and the ability to incorporate sub-setting within the masking process. Note that the use of Simple Data Masking does not preclude any of these functions; simply that you would have to do them manually or through the use of some additional (possibly third party) tool. It should be clear from this that Fast Data Masking is significantly richer than Simple Data Masking. One would therefore wish that the former was available for use with a much broader range of sources than is currently the case. Grid-Tools is very pragmatic about this: if there is appropriate demand for Fast Data Masking for a particular data source then they will build the relevant connector. File-based masking is often required in conjunction with database masking. For example, you might have scrambled the social security numbers in the target database. However, your input file could now contain non-matching social security numbers and the load will fail. Using the cross-reference table, or the hash routines employed by Grid-Tools as a part of the masking process, you can ensure that this mismatch doesn t happen. In practice, in terms of the actual steps used in file-based masking, you register the file definitions, identify (profile) the internal structure of the file and its relationships, import a sample file to make sure that the file definition and sample file match, define any sensitivity and any data manipulation functions, and then run the scramble utility. As an example of the interface used for this process, Figure 2 illustrates how this works along with the general look and feel of the Datamaker product. Figure 2: The Datamaker interface: marking sensitive table columns Bloor Research A Bloor InDetail Paper
9 Product details Finally, it is worth noting that there is an emerging distinction between static data masking, which is used in test and development environments and which we have largely been describing, and dynamic data masking. The latter is used in conjunction with operational data in real-time. As mentioned, Grid- Tools products can be deployed in a dynamic as well as a static fashion. Sub-setting There are several approaches provided for sub-setting but basically you define a driving table, a view or a temporary table (typically a subset of a driving table). Assuming a driving table, this will be extended along with any related tables defined by a primary-foreign key relationship (though these can also be selectively removed) and you can then add any others that might be linked in application terms, by defining a relevant join condition. You can view the data as you design your subsets by means of a graphical user interface. Features include the ability to shrink a database in situ so that you can reorganise your tables into different tablespaces and compress you data; the ability to validate data quality during testing; to compare data before and after a transaction, regardless of whether changes were expected or not; and date shift capabilities that will keep time-sensitive data up-to-date. The actual process of sub-setting uses the native capabilities of the relevant data source such as Oracle Data Pump, DB2 import/export and so on. Coverage Once you have sub-setted and masked your data, thought should be given to coverage. The same will also apply if generating completely synthetic data. Coverage in Datamaker is provided through the Data Design module. This is, in fact, BenderRTM (formerly CaliberRTM), which has been OEM d by Grid-Tools, integrated into its environment, and extended, most notably by the use of Spotfire (TIBCO) technology for visualising coverage maps. Data Design actually offers two test design engines: Quick Design, which is based on orthogonal or optimised pairs; and Cause-Effect Graphing, which is the more advanced of the two engines. Figure 3 shows the graphical design interface that you use in conjunction with the latter engine. As can be seen, you use Figure 3: The Datamaker interface for Cause-Effect Graphing A Bloor InDetail Paper Bloor Research
10 Product details this to define relationships, constraints and attributes and then the software will generate relevant test cases based on the design you have created. Options include the ability to re-run old tests, to generate new test cases from the same diagram, and there are also facilities to link in any existing test cases that may have been created outside of the environment. Reports generated include a Functional Variation report (which will highlight any logical inconsistencies) and a Script Test Definition report as well as various matrices and the aforementioned coverage map. Synthetic data The basic principle behind the generation of synthetic data is that you discover your data structures, define the characteristics of the data you want to generate, generate the synthetic data and then store the results in what Grid-Tools calls a Test Data Warehouse, which is shown in Figure 4. Once you have a data model that the generated data needs to match, as discussed previously, you now need to define the characteristics of the synthetic data. At a mundane level you can select the assignation of attributes: for example, what percentage of female first names versus males, which can then be generated from seed tables. Then you can define what percentage of errors you want in the data and what sorts of errors. You can also, using the product s Data Shaper module, identify trends and patterns in the existing data that you may want to include, or expand upon, in the synthetic data. For instance, using Data Shaper you might have identified that the shape of your data is gradually changing (perhaps because your online business is growing rapidly and typical order sizes and values are different from conventional retail sales) and you might want to project that, going forward, for testing purposes. Data Shaper has a large number of parametric tests it can use for identifying such trends, including Weibull, Pareto, Chi-Squared, Gaussian Bell-curve and other tests. There is also support for time series analysis though, unlike the other tests, this is not currently automated (the company is working on this). Once those definitions are complete you can assemble these objects (an object consisting of data together with relevant transformations and selection criteria) into scenarios, where a scenario might represent customers and their interactions, or products and their relationships, or other types of business entity. Scenarios may be arranged hierarchically so that one scenario can inherit characteristics from another. You manipulate scenarios via a graphical interface (portal). Also within this portal you can perform coverage analysis, as discussed previously. Figure 4: Datamaker architecture Bloor Research A Bloor InDetail Paper
11 Product details Once you are happy with the preceding processes you select the relevant scenario and generate the data for it. Note that Datamaker has transformation capabilities (similar to those within an ETL tool). The results are stored in the Test Data Warehouse and may be re-generated on demand. This is a really important feature. During development, specifications change. This often means that you need to change the dataset that you are testing against, which can be an onerous process in conventional environments. With Datamaker you simply change the relevant attributes within your scenario and re-generate the data. In agile development environments, in particular, this means that the test data can be as agile as the development and application testing, which is not usually the case. Full version control is maintained by the Test Data Warehouse, so that you can go back to previous versions if necessary. One final point that is worthy of mention is that Datamaker is not limited to generating database-specific data. It can also be used, for example, to populate the spreadsheets that are typically used with tools such as HP s QuickTest Professional and LoadRunner products. SOA Data Pro SOA Data Pro is used to provide the back-end data needed to service SOA requests so that testing becomes more automated and less manual. It includes a repository that allows you to store and freeze generated test data, in the former case so that this can be referenced by downstream requests and in the latter case to support the testing of multiple SOA requests. You can also store spreadsheets within the repository, thereby providing a central point of control. SOA Data Pro integrates with products like soapui Pro from Eviware, SOA Test from ParaSoft, HP s Service Test and itko s LISA. Data Archive Data Archive is a stand-alone product that is not part of Datamaker but we include it here for the sake of completeness. For archival purposes you need to understand your data in much the same way as you need to for generating synthetic data and Data Archive reuses the capabilities already described in Datamaker for that purpose. Similarly, the product reuses Datamaker s sub-setting capabilities for extracting the data to be archived, along with join views that span the archived and live data. The software will generate delete scripts to support purging and, of course, there are masking options. The product integrates with various third party archival vendors including Bridgehead, Mobius (ASG) and SAND Technology. A Bloor InDetail Paper Bloor Research
12 The vendor Grid-Tools was founded in 2004 though, in a sense, its foundations go further back than that, since its founders had previously built up and then sold BitByBit to OuterBay (since acquired by HP), so the company has a depth of experience in this area. This also explains why the company is privately owned and self-financing. The company s headquarters are in the UK and it also has offices in the United States and India. It has an extensive partner programme with trained and certified staff covering Australia, Austria, Belgium, Canada, France, Germany, Ireland, Israel, Italy, Latin America, New Zealand, Portugal, Scandinavia, Singapore, South Africa, South Korea, Spain, Switzerland and the Netherlands as well as the countries in which it maintains offices (where it also has partners). Partners tend to be consulting houses and systems integrators, both local and international. The latter category includes Cap Gemini, Cognizant, EDS, Infosys, Birlasoft, SQS and CSC amongst others. On the technical side, Grid-Tools has partnerships with MySQL (Oracle), Oracle (which is also a customer), Compuware, Bridgehead Software, InterSystems, Silwood Technology, SAND Technology and various specialist testing vendors. The company also OEMs technology from Spotfire (TIBCO), Bender and Pervasive. Web site: Bloor Research A Bloor InDetail Paper
13 Summary In our view, Datamaker is the most extensive and most complete test data management product that is available on the market today. In particular, it can do things that other players in the market cannot. On the downside, Grid-Tools is a relatively small organisation in a market that is dominated by major vendors. In that sort of scenario the only way to stay above water is to be technically more advanced as well as generally being smarter and more agile. That is exactly what Grid-Tools is doing and there is every indication that it will continue to do so. Further Information Further information about this subject is available from A Bloor InDetail Paper Bloor Research
14 Bloor Research overview Bloor Research is one of Europe s leading IT research, analysis and consultancy organisations. We explain how to bring greater Agility to corporate IT systems through the effective governance, management and leverage of Information. We have built a reputation for telling the right story with independent, intelligent, well-articulated communications content and publications on all aspects of the ICT industry. We believe the objective of telling the right story is to: Describe the technology in context to its business value and the other systems and processes it interacts with. Understand how new and innovative technologies fit in with existing ICT investments. Look at the whole market and explain all the solutions available and how they can be more effectively evaluated. Filter noise and make it easier to find the additional information or news that supports both investment and implementation. Ensure all our content is available through the most appropriate channel. Founded in 1989, we have spent over two decades distributing research and analysis to IT user and vendor organisations throughout the world via online subscriptions, tailored research services, events and consultancy projects. We are committed to turning our knowledge into business value for you. About the author Philip Howard Research Director - Data Philip started in the computer industry way back in 1973 and has variously worked as a systems analyst, programmer and salesperson, as well as in marketing and product management, for a variety of companies including GEC Marconi, GPT, Philips Data Systems, Raytheon and NCR. After a quarter of a century of not being his own boss Philip set up what is now P3ST (Wordsmiths) Ltd in 1992 and his first client was Bloor Research (then ButlerBloor), with Philip working for the company as an associate analyst. His relationship with Bloor Research has continued since that time and he is now Research Director. His practice area encompasses anything to do with data and content and he has five further analysts working with him in this area. While maintaining an overview of the whole space Philip himself specialises in databases, data management, data integration, data quality, data federation, master data management, data governance and data warehousing. He also has an interest in event stream/complex event processing. In addition to the numerous reports Philip has written on behalf of Bloor Research, Philip also contributes regularly to and Analysis.com and was previously the editor of both Application Development News and Operating System News on behalf of Cambridge Market Intelligence (CMI). He has also contributed to various magazines and published a number of reports published by companies such as CMI and The Financial Times. Away from work, Philip s primary leisure activities are canal boats, skiing, playing Bridge (at which he is a Life Master) and walking the dog.
15 Copyright & disclaimer This document is copyright 2011 Bloor Research. No part of this publication may be reproduced by any method whatsoever without the prior consent of Bloor Research. Due to the nature of this material, numerous hardware and software products have been mentioned by name. In the majority, if not all, of the cases, these product names are claimed as trademarks by the companies that manufacture the products. It is not Bloor Research s intent to claim these names or trademarks as our own. Likewise, company logos, graphics or screen shots have been reproduced with the consent of the owner and are subject to that owner s copyright. Whilst every care has been taken in the preparation of this document to ensure that the information is correct, the publishers cannot accept responsibility for any errors or omissions.
16 2nd Floor, St John Street LONDON, EC1V 4PY, United Kingdom Tel: +44 (0) Fax: +44 (0) Web:
White Paper. Lower your risk with application data migration. next steps with Informatica
White Paper Lower your risk with application data migration A White Paper by Bloor Research Author : Philip Howard Publish date : April 2013 If we add in Data Validation and Proactive Monitoring then Informatica
White Paper. Data Migration
White Paper Data Migration A White Paper by Bloor Research Author : Philip Howard Publish date : May 2011 data migration projects are undertaken because they will support business objectives. There are
White Paper. SAP ASE Total Cost of Ownership. A comparison to Oracle
White Paper SAP ASE Total Cost of Ownership A White Paper by Bloor Research Author : Philip Howard Publish date : April 2014 The results of this survey are unequivocal: for all 21 TCO and related metrics
Sagent Data Flow. from Group 1 Software. an extract from the Bloor Research report, Data Integration, Volume 1
Sagent Data Flow from Group 1 Software an extract from the Bloor Research report, Data Integration, Volume 1 Sagent Data Flow Sagent Data Flow Fast facts Sagent Data Flow, which is now provided by Group
Data Integration Platforms - Talend
Data Integration Platforms - Talend Author : Philip Howard Publish date : July 2008 page 1 Introduction Talend is an open source provider of data integration products. However, while many open source
White Paper. What the ideal cloud-based web security service should provide. the tools and services to look for
White Paper What the ideal cloud-based web security service should provide A White Paper by Bloor Research Author : Fran Howarth Publish date : February 2010 The components required of an effective web
IBM InfoSphere Optim Test Data Management solution for Oracle E-Business Suite
IBM InfoSphere Optim Test Data Management solution for Oracle E-Business Suite Streamline test-data management and deliver reliable application upgrades and enhancements Highlights Apply test-data management
InDetail. Kdb+ and the Internet of Things/Big Data
InDetail Kdb+ and the Internet of Things/Big Data An InDetail Paper by Bloor Research Author : Philip Howard Publish date : August 2014 Kdb+ has proved itself in what is unarguably the most demanding big
IBM InfoSphere Optim Test Data Management Solution
IBM InfoSphere Optim Test Data Management Solution Highlights Create referentially intact, right-sized test databases Automate test result comparisons to identify hidden errors Easily refresh and maintain
ORACLE DATA INTEGRATOR ENTEPRISE EDITION FOR BUSINESS INTELLIGENCE
ORACLE DATA INTEGRATOR ENTEPRISE EDITION FOR BUSINESS INTELLIGENCE KEY FEATURES AND BENEFITS (E-LT architecture delivers highest performance. Integrated metadata for alignment between Business Intelligence
IBM InfoSphere Optim Test Data Management
IBM InfoSphere Optim Test Data Management Highlights Create referentially intact, right-sized test databases or data warehouses Automate test result comparisons to identify hidden errors and correct defects
Test Data Management. Services Catalog
Test Data Management Services Catalog TABLE OF CONTENTS WHY TDM?... 3 CLIENT SUCCESS STORIES... 4 Large Automotive Company Data Generation...4 Large Entertainment Company Coverage Analysis...4 TDM APPROACH...
A WHITE PAPER By Silwood Technology Limited
A WHITE PAPER By Silwood Technology Limited Using Safyr to facilitate metadata transparency and communication in major Enterprise Applications Executive Summary Enterprise systems packages such as SAP,
ORACLE DATA INTEGRATOR ENTERPRISE EDITION
ORACLE DATA INTEGRATOR ENTERPRISE EDITION ORACLE DATA INTEGRATOR ENTERPRISE EDITION KEY FEATURES Out-of-box integration with databases, ERPs, CRMs, B2B systems, flat files, XML data, LDAP, JDBC, ODBC Knowledge
Sybase Solutions for Healthcare Adapting to an Evolving Business and Regulatory Environment
Sybase Solutions for Healthcare Adapting to an Evolving Business and Regulatory Environment OVERVIEW Sybase Solutions for Healthcare Adapting to an Evolving Business and Regulatory Environment Rising medical
Test Data Management Concepts
Test Data Management Concepts BIZDATAX IS AN EKOBIT BRAND Executive Summary Test Data Management (TDM), as a part of the quality assurance (QA) process is more than ever in the focus among IT organizations
ENTERPRISE EDITION ORACLE DATA SHEET KEY FEATURES AND BENEFITS ORACLE DATA INTEGRATOR
ORACLE DATA INTEGRATOR ENTERPRISE EDITION KEY FEATURES AND BENEFITS ORACLE DATA INTEGRATOR ENTERPRISE EDITION OFFERS LEADING PERFORMANCE, IMPROVED PRODUCTIVITY, FLEXIBILITY AND LOWEST TOTAL COST OF OWNERSHIP
InDetail. RainStor archiving
InDetail RainStor archiving An InDetail Paper by Bloor Research Author : Philip Howard Publish date : November 2013 Archival is a no-brainer when it comes to return on investment and total cost of ownership.
IBM InfoSphere Optim Data Masking solution
IBM InfoSphere Optim Data Masking solution Mask data on demand to protect privacy across the enterprise Highlights: Safeguard personally identifiable information, trade secrets, financials and other sensitive
An Oracle White Paper March 2012. Managing Metadata with Oracle Data Integrator
An Oracle White Paper March 2012 Managing Metadata with Oracle Data Integrator Introduction Metadata information that describes data is the foundation of all information management initiatives aimed at
The ESB and Microsoft BI
Business Intelligence The ESB and Microsoft BI The role of the Enterprise Service Bus in Microsoft s BI Framework Gijsbert Gijs in t Veld CTO, BizTalk Server MVP [email protected] About motion10
Datamaker for Skytap. Provide full-sized environments filled with up-to-date test data in minutes
Datamaker for Skytap Provide full-sized environments filled with up-to-date test data in minutes Is your testing constrained by environments and data? As applications have become more complex, provisioning
Datamaker - the complete Test Data Management solution
Datamaker - the complete Test Data Management solution Improve software application quality whilst reducing time-to-market Executive Summary Whether building and testing new applications, re-engineering
An Oracle White Paper February 2014. Oracle Data Integrator 12c Architecture Overview
An Oracle White Paper February 2014 Oracle Data Integrator 12c Introduction Oracle Data Integrator (ODI) 12c is built on several components all working together around a centralized metadata repository.
Managing Third Party Databases and Building Your Data Warehouse
Managing Third Party Databases and Building Your Data Warehouse By Gary Smith Software Consultant Embarcadero Technologies Tech Note INTRODUCTION It s a recurring theme. Companies are continually faced
Real-Time Data Access Using Restful Framework for Multi-Platform Data Warehouse Environment
www.wipro.com Real-Time Data Access Using Restful Framework for Multi-Platform Data Warehouse Environment Pon Prabakaran Shanmugam, Principal Consultant, Wipro Analytics practice Table of Contents 03...Abstract
Getting started with API testing
Technical white paper Getting started with API testing Test all layers of your composite applications, not just the GUI Table of contents Executive summary... 3 Introduction... 3 Who should read this document?...
White Paper. Thirsting for Insight? Quench It With 5 Data Management for Analytics Best Practices.
White Paper Thirsting for Insight? Quench It With 5 Data Management for Analytics Best Practices. Contents Data Management: Why It s So Essential... 1 The Basics of Data Preparation... 1 1: Simplify Access
White Paper. Architecting the security of the next-generation data center. why security needs to be a key component early in the design phase
White Paper Architecting the security of the next-generation data center A White Paper by Bloor Research Author : Fran Howarth Publish date : August 2011 teams involved in modernization projects need to
Optimizing the Performance of the Oracle BI Applications using Oracle Datawarehousing Features and Oracle DAC 10.1.3.4.1
Optimizing the Performance of the Oracle BI Applications using Oracle Datawarehousing Features and Oracle DAC 10.1.3.4.1 Mark Rittman, Director, Rittman Mead Consulting for Collaborate 09, Florida, USA,
How to address top problems in test data management
How to address top problems in test data management Data reuse, sub-setting and masking Business white paper Table of contents Why you need test data management... 3 The challenges of preparing and managing
Buying vs. Building Business Analytics. A decision resource for technology and product teams
Buying vs. Building Business Analytics A decision resource for technology and product teams Introduction Providing analytics functionality to your end users can create a number of benefits. Actionable
SAP BusinessObjects Business Intelligence 4.1 One Strategy for Enterprise BI. May 2013
SAP BusinessObjects Business Intelligence 4.1 One Strategy for Enterprise BI May 2013 SAP s Strategic Focus on Business Intelligence Core Self-service Mobile Extreme Social Core for innovation Complete
Enabling Better Business Intelligence and Information Architecture With SAP Sybase PowerDesigner Software
SAP Technology Enabling Better Business Intelligence and Information Architecture With SAP Sybase PowerDesigner Software Table of Contents 4 Seeing the Big Picture with a 360-Degree View Gaining Efficiencies
Test Data Management Best Practice
Test Data Management Best Practice, Inc. 5210 Belfort Parkway, Suite 400 Author: Stephanie Chace Quality Practice Lead [email protected], Inc. 2011 www.meridiantechnologies.net Table of
DBKDA 2012 : The Fourth International Conference on Advances in Databases, Knowledge, and Data Applications
Evaluation of Data Anonymization Tools Sergey Vinogradov Corporate Technology Siemens LLC Saint-Petersburg, Russia [email protected] Alexander Pastsyak Corporate Technology Siemens LLC Saint-Petersburg,
Integrating data in the Information System An Open Source approach
WHITE PAPER Integrating data in the Information System An Open Source approach Table of Contents Most IT Deployments Require Integration... 3 Scenario 1: Data Migration... 4 Scenario 2: e-business Application
534441 IPI*GrammTech Original Price List 11.25.13
Informatica 0000001313-0001 PowerExchange for Flat Files - Batch Option - (1-3 Servers) per Server Production each $ 16,458.44 Informatica 0000001314-0001 PowerExchange for Flat Files - Batch Option -
Genio. from Hummingbird. an extract from the Bloor Research report, Data Integration, Volume 1
Genio from Hummingbird an extract from the Bloor Research report, Data Integration, Volume 1 Hummingbird Genio Hummingbird Genio Fast facts Hummingbird s Integration Suite component, Genio, is a data
High-Volume Data Warehousing in Centerprise. Product Datasheet
High-Volume Data Warehousing in Centerprise Product Datasheet Table of Contents Overview 3 Data Complexity 3 Data Quality 3 Speed and Scalability 3 Centerprise Data Warehouse Features 4 ETL in a Unified
A Road Map to Successful Data Masking
A Road Map to Successful Data Masking Secure your production data By Huw Price, Managing Director 2 Introduction How valuable is data masking to your organisation? According to the Ponemon Institute s
TECHNOLOGY BRIEF: CA ERWIN SAPHIR OPTION. CA ERwin Saphir Option
TECHNOLOGY BRIEF: CA ERWIN SAPHIR OPTION CA ERwin Saphir Option Table of Contents Executive Summary SECTION 1 2 Introduction SECTION 2: OPPORTUNITY 2 Modeling ERP Systems ERP Systems and Data Warehouses
Data processing goes big
Test report: Integration Big Data Edition Data processing goes big Dr. Götz Güttich Integration is a powerful set of tools to access, transform, move and synchronize data. With more than 450 connectors,
Build a Streamlined Data Refinery. An enterprise solution for blended data that is governed, analytics-ready, and on-demand
Build a Streamlined Data Refinery An enterprise solution for blended data that is governed, analytics-ready, and on-demand Introduction As the volume and variety of data has exploded in recent years, putting
Enabling Better Business Intelligence and Information Architecture With SAP PowerDesigner Software
SAP Technology Enabling Better Business Intelligence and Information Architecture With SAP PowerDesigner Software Table of Contents 4 Seeing the Big Picture with a 360-Degree View Gaining Efficiencies
ORACLE BUSINESS INTELLIGENCE, ORACLE DATABASE, AND EXADATA INTEGRATION
ORACLE BUSINESS INTELLIGENCE, ORACLE DATABASE, AND EXADATA INTEGRATION EXECUTIVE SUMMARY Oracle business intelligence solutions are complete, open, and integrated. Key components of Oracle business intelligence
Accurate identification and maintenance. unique customer profiles are critical to the success of Oracle CRM implementations.
Maintaining Unique Customer Profile for Oracle CRM Implementations By Anand Kanakagiri Editor s Note: It is a fairly common business practice for organizations to have customer data in several systems.
IBM Cognos 8 Business Intelligence Analysis Discover the factors driving business performance
Data Sheet IBM Cognos 8 Business Intelligence Analysis Discover the factors driving business performance Overview Multidimensional analysis is a powerful means of extracting maximum value from your corporate
Real-Time Database Protection and. Overview. 2010 IBM Corporation
Real-Time Database Protection and Monitoring: IBM InfoSphere Guardium Overview Agenda Business drivers for database security InfoSphere Guardium architecture Common applications The InfoSphere portfolio
PATROL From a Database Administrator s Perspective
PATROL From a Database Administrator s Perspective September 28, 2001 Author: Cindy Bean Senior Software Consultant BMC Software, Inc. 3/4/02 2 Table of Contents Introduction 5 Database Administrator Tasks
Database-Archiving Products Are Gaining Market Traction
Database-Archiving Products Are Gaining Market Traction Gartner RAS Core Research Note G00161796, Carolyn DiCenzo, 13 October 2008 Few vendors are offering solutions for archiving data from custom-built
OpenText Actuate Big Data Analytics 5.2
OpenText Actuate Big Data Analytics 5.2 OpenText Actuate Big Data Analytics 5.2 introduces several improvements that make the product more useful, powerful and flexible for end users. A new data loading
Implementing efficient system i data integration within your SOA. The Right Time for Real-Time
Implementing efficient system i data integration within your SOA The Right Time for Real-Time Do your operations run 24 hours a day? What happens in case of a disaster? Are you under pressure to protect
JRules 6.0. from ILOG
JRules 6.0 from ILOG Fast facts ILOG JRules puts the rules that define the business back into the hands of the business. With traditional application development, business rules have been implemented deep
data express DATA SHEET OVERVIEW
data express DATA SHEET OVERVIEW The reliability of IT systems is a key requirement of almost any organization. Unexpected failure of enterprise systems can be expensive and damaging to an organization.
DATA MASKING A WHITE PAPER BY K2VIEW. ABSTRACT K2VIEW DATA MASKING
DATA MASKING A WHITE PAPER BY K2VIEW. ABSTRACT In today s world, data breaches are continually making the headlines. Sony Pictures, JP Morgan Chase, ebay, Target, Home Depot just to name a few have all
Database-driven library system
Database-driven library system Key-Benefits of CADSTAR 12.1 Characteristics of database-driven library system KEY-BENEFITS Increased speed when searching for parts You can edit/save a single part (instead
Data Masking Checklist
Data Masking Checklist Selecting the Right Data Masking Tool Selecting Your Masking Tool Ensuring compliance with current data protection regulations and guidelines has become a mandatory operation. Non-compliance
Business Intelligence and Service Oriented Architectures. An Oracle White Paper May 2007
Business Intelligence and Service Oriented Architectures An Oracle White Paper May 2007 Note: The following is intended to outline our general product direction. It is intended for information purposes
Integrated Data Management: Discovering what you may not know
Integrated Data Management: Discovering what you may not know Eric Naiburg [email protected] Agenda Discovering existing data assets is hard What is Discovery Discovery and archiving Discovery, test
Is ETL Becoming Obsolete?
Is ETL Becoming Obsolete? Why a Business-Rules-Driven E-LT Architecture is Better Sunopsis. All rights reserved. The information contained in this document does not constitute a contractual agreement with
The Ultimate Guide to Buying Business Analytics
The Ultimate Guide to Buying Business Analytics How to Evaluate a BI Solution for Your Small or Medium Sized Business: What Questions to Ask and What to Look For Copyright 2012 Pentaho Corporation. Redistribution
COSCON: SAP Technologies Help Create a Unified Enterprise Data Platform for Global Operations
2013 SAP AG or an SAP affiliate company. All rights reserved. COSCON: SAP Technologies Help Create a Unified Enterprise Data Platform for Global Operations COSCON Container Lines Co. Ltd. Industry Transportation
Why Buy GoldMine Premium Edition 9.2?
Why Buy GoldMine Premium Edition 9.2? The Top Reasons Why to Buy GoldMine Premium Edition with New Items in Version 9.2 GoldMine is a leading provider of relationship management solutions for small businesses
Selecting the Right Change Management Solution Key Factors to Consider When Evaluating Change Management Tools for Your Databases and Teams
Tech Notes Selecting the Right Change Management Solution Key Factors to Consider When Evaluating Change Management Tools for Your Databases and Teams Embarcadero Technologies July 2007 Corporate Headquarters
Oracle Business Intelligence Publisher. 1 Oracle Business Intelligence Publisher Certification. Certification Information 10g Release 3 (10.1.3.4.
Oracle Business Intelligence Publisher Certification Information 10g Release 3 (10.1.3.4.2) E12692-08 September 2011 This document outlines the certified hardware and software configurations for Oracle
CRM for Customer Service and Support
CRM for Customer Service and Support Maximize satisfaction. Drive repeat business Servicing customers effectively is a key component in growing your business through loyal, repeat buyers. Whether it s
The Ultimate Guide to Buying Business Analytics
The Ultimate Guide to Buying Business Analytics How to Evaluate a BI Solution for Your Small or Medium Sized Business: What Questions to Ask and What to Look For Copyright 2012 Pentaho Corporation. Redistribution
Business-driven governance: Managing policies for data retention
August 2013 Business-driven governance: Managing policies for data retention Establish and support enterprise data retention policies for ENTER» Table of contents 3 4 5 Step 1: Identify the complete business
Lawson Talent Management
Lawson Talent Imagine Knowing: Which employees have the highest potential how to recruit more talent like them. Imagine Understanding: Which employees are a flight risk how your compensation plans could
JD Edwards EnterpriseOne Tools. 1 Understanding JD Edwards EnterpriseOne Business Intelligence Integration. 1.1 Oracle Business Intelligence
JD Edwards EnterpriseOne Tools Embedded Business Intelligence for JD Edwards EnterpriseOne Release 8.98 Update 4 E21426-02 March 2011 This document provides instructions for using Form Design Aid to create
HOW INTERSYSTEMS TECHNOLOGY ENABLES BUSINESS INTELLIGENCE SOLUTIONS
HOW INTERSYSTEMS TECHNOLOGY ENABLES BUSINESS INTELLIGENCE SOLUTIONS A white paper by: Dr. Mark Massias Senior Sales Engineer InterSystems Corporation HOW INTERSYSTEMS TECHNOLOGY ENABLES BUSINESS INTELLIGENCE
IBM AND NEXT GENERATION ARCHITECTURE FOR BIG DATA & ANALYTICS!
The Bloor Group IBM AND NEXT GENERATION ARCHITECTURE FOR BIG DATA & ANALYTICS VENDOR PROFILE The IBM Big Data Landscape IBM can legitimately claim to have been involved in Big Data and to have a much broader
