Making data predictive why reactive just isn t enough
|
|
|
- Ariel Cooper
- 9 years ago
- Views:
Transcription
1 Making data predictive why reactive just isn t enough Andrew Peterson, Ph.D. Principal Data Scientist Soltius NZ, Ltd. New Zealand 2014 Big Data and Analytics Forum 18 August, 2014 Caveats and disclaimer: This is a large and complex topic and I cannot do it justice in the time available, so I will focus on my own perspectives (i.e., opinions) and experience. As such, I am not presenting or representing the views or opinions of any current or past company or organisation I may have worked with or for. I'm also going to brush aside all technical details, despite how important they really are. So please keep that in mind if I say something you disagree with. We can chat about details later if that would be helpful for you. Introduction When we talk about predictive analytics, most people probably think about using computers to predict an event or the state of a system at some future point in time. But there are other types of predictions that we can make using analytics that don't necessarily involve time directly although they may indirectly. These include things like estimating the output of a production line if certain control variables are changed, understanding how a change in hospital staffing levels may affect patient occupancy, or estimating the risk of a person developing a specific heritable disease. Answers to these types of problems are predictive because they involve estimating the state of a system under conditions that have not yet been observed, even though they may not necessarily involve explicitly projecting the state of a system forward in time. Going beyond analysing past usage data to estimate future trends In my opinion, and in very general terms, the principal feature that distinguishes predictive analytics from nonpredictive analytics is statistical and mathematical modelling. Consequently, to move from non-predictive analytics to predictive analytics requires at least the following two steps: 1. The development and use of a statistical or mathematical model to generalise historical patterns among input variables and output variables for the system under study. 2. The use of that model to estimate a new, unobserved output state from a set of at least partly unobserved input states. The principal requirement for predictive analytics is therefore a model, that is, a mathematical representation of the nature and the magnitude of relationships among the input and output variables. The model itself is an abstraction of reality. But if the system under study is amenable to predictive modelling (many systems are not), and the modelling is done well, then the model will enhance our understanding of the system through its representation of the relationships among variables. And with enhanced understanding comes greater opportunities for effective management and control of the target system. In other words, enhanced understanding empowers us to make better decisions. This, to me, is the real benefit that comes from predictive analytics.
2 Benefits of looking forward developing frameworks for better business decisions As previously mentioned, one of the most important benefits that predictive analytics can provide when it is done well is to assist organisations in making better management and control decisions. Put slightly differently, good predictive models can help organisations make important business and operational decisions proactively instead of reactively. But the true business value of being able to do this can only be assessed on a case-by-case basis and requires a thorough analysis of an organisation's business needs and opportunities. The principal reason most organisations will consider engaging in predictive analytics is the presumption that it will enhance the organisation's financial position either through increasing revenue or reducing costs. Consequently, an organisation must be able to rationally and unemotionally evaluate the true costs, the realistic benefits, and the potential risk of failure before implementing a programme in predictive analytics. Incorporating predictive analytics into a framework for making better business decisions requires the information provided by the models to be disseminated effectively to those who can use it to help the organisation achieve its strategic goals. Therefore, an organisation will need a comprehensive technological, strategic, and operational framework for deploying predictive models and for disseminating the insight they generate. Exactly what this framework should look like will depend on the organisation, its key business, and its objectives. But its importance should not be underestimated because even the best predictive model will be next to useless if it cannot empower people or systems to work more effectively. Predictive modelling pitfalls, challenges, and integration with existing enterprise software The following is a non-exhaustive list of issues I would advise any organisation to consider in detail before investing in predictive analytics. Some of these issues will have more or less relevance to different organisations, but I present them here because I have seen organisations, teams, and individuals trip over them in the past (and sometimes spectacularly). Pitfalls 1. Failing to understand the true costs, the realistic benefits, the risk of failure, and the likely impact each of these may have on the organisation: Costs of implementing predictive analytics often go well beyond software and hardware and can include training costs, costs associated with low productivity as staff learn new technologies and programming languages, costs associated with developing production-ready models, costs associated with maintaining and retuning models over time, etc. Understanding the realistic benefits of predictive analytics requires engaging people with substantive experience in the discipline who can evaluate a given business problem and the data that's available in order to provide an informed opinion of the likelihood of success. Some applications of predictive analytics are so well established and so general across different business sectors (e.g. some market analysis models) that identifying realistic benefits is relatively straight forward. But the more innovation that's required in building a predictive model, the less certain the outcome will be. Simply put, some systems can't be modelled effectively at least to the extent that an organisation could benefit from the effort. Consequently, an organisation must understand the impact that failure may have not only in terms of direct costs, but also in terms of the organisations strategic goals, etc. 2. Failing to employ or engage people who truly understand the domain: Predictive analytics is a specialised discipline and the devil really is in the details. It is very easy to do predictive analytics
3 poorly and the cost of doing so can be very high. Candidates for this type of work should have a comprehensive understanding of concepts such as: a. Statistical independence b. Autocorrelation c. Multicolinearity d. Heteroscedasticity e. Pseudoreplication f. Nonstationarity g. Statistical power h. Spurious regression i. Conditional probabilities j. Over fitting k. Etc., etc Data snooping the curse of data mining: Data snooping can take many forms but in essence it means poking around data and making many different comparisons until some interesting relationship is found, then reporting that relationship as if it is real and meaningful. But, unless there has been appropriate statistical control or validation against independent data, it's highly probable that the "interesting" relationship may be spurious and simply the outcome of random chance. To illustrate, I generated 100 random time series using an ARIMA (1,1,0) model with an AR term of 0.1 (examples are shown in the figure below). The key point to understand here is that each separate time series has no causal relationship with any other time series in the set, and that each series wanders up and down randomly in a manner similar to some stock prices. The only thing they have in common is their starting value and the statistical properties of their random wandering. If we calculate all 4950 pairwise correlations between these series, then roughly half of the correlations will be greater than ± 0.5. Most analysts would probably agree that a correlation greater than ± 0.5 should warrant further investigation. However, we already know that all the series in this example are random and independent of each other. Yet our analysis has found roughly 2000 pairs of variables that suggest potentially interesting, if not important, relationships. Even though this is a deliberate and contrived example, it nevertheless illustrates the point that the more variables we have in our data and the more relationships we investigate, the more we need to expect that we will find spurious relationships that have no causal or business basis. The take-home message here is to be wary of those who emphasise the simplicity of finding "hidden relationships and opportunities in masses of data" without detailing the precautions they take to ensure those relationships are real.
4 4. Forgetting the "Science" in "Data Science": In my opinion, the term "Science" appears in the title "Data Science" for a very good reason to emphasise the need to take a rigorous, hypothesis driven approach to system modelling and data analysis. Such an approach would avoid the data snooping problem just mentioned because a Data Scientist would work with an organisation's subject matter experts to develop a range of hypotheses about the business problem under study. In addition, the Data Scientist would develop appropriate modelling and statistical frameworks for testing those hypotheses before running any type of analysis. Having said that, it is appropriate to "poke around" in data when certain conditions are met, such as segregating data into multiple sets where one set is used only for exploratory analysis, another set is used only for model development and validation, and a third set is used only for model testing. Other approaches can include using strict statistical criteria for making multiple comparisons, etc. The critical point though is that building a valid predictive model that can return business value in a production setting is not straightforward and requires the rigorous application of the scientific method in order to maximise success. 5. Overloading on expensive and/or complex technology that may not be necessary. This can be avoided through informed, rational, and pragmatic choices that constrain costs while providing sufficient flexibility for projected growth. Challenges 1. Data quality vs. quantity: This is paramount if you can't trust your data then it doesn't matter how much of it you have because you will never be able to trust any forecasts derived from it. However, data quality generally follows a continuum from useful to useless, or good to bad. But how "good" does your data need to be in order to answer the questions that are important to your organisation? This can be addressed through careful simulation and is often referred to as a Power Analysis in statistics. That is, given data with a certain amount of variability, how large does our sample need to be in order to identify an effect of a certain size and with a specific level of confidence when that effect really does exist? There is, however, a pragmatic flip-side to the data quality issue: Is it better to have an approximate
5 answer to an important question or a precise answer to a trivial one? Can we make an informed trade-off between data quality and business value? A robust business case for predictive analytics should address these types of questions. 2. Controlled Experiments vs. Observational Studies: Models based on controlled experiments will always provide more understanding about a system than models based on observational data from the same system. But controlled experiments may be difficult, expensive, or even impossible to perform in many situations. If a controlled experiment can't be done, then data quality and the design and interpretation of the analysis become even more important. 3. Big Data vs. sufficient data: a. Is more data really better? To the best of my understanding there is no statistical theory that shows more data is always better, as long as we have sufficient data for the problem at hand. b. Statistical significance vs. practical significance: Sufficient data means having enough data to be able to identify patterns and trends that will have real business implications. In general, the more data we have then the greater our ability to identify smaller and smaller patterns in our data if those patterns truly exist. But at what point does a pattern or trend become so small that it has no practical business implications at all, even if it is statistically significant? And why would we use the volumes of data required to identify very small patterns if they have no practical significance for the business? Again, a robust business case for predictive analytics should address this type of question. It also leads directly to the big old elephant lounging in the Big Data Analytics room that many people seem to ignore computation time. c. Computation Time: Fitting predictive models to data takes computer time, and the more data we use in those models the longer the time required to fit them. Identifying the best model for a given data set requires fitting the model many time often thousands of times! For example, a support vector machine may have 3 tuning parameters. We need to find the best combination of those parameters for our data. We decide to test all combinations of 5 different values for each tuning parameter to give a total of 125 different parameter combinations. We want a good estimate of the model error because it will be used to make important business decisions, so we use a 10-fold cross-validation, repeated 3 times, for each of the 125 parameter combinations to give a total of 3750 separate model fittings. We also thought it best to use 5 million records in or data file even though a power analysis estimated that 500,000 records would be sufficient for our business requirements. With 5 million records it may take 5 minutes to fit the model once using a single core on our server. This would result in a total fitting time of hours, or roughly 13 days. But we are smart and have good software and we know how to parallelise the solution. With 13 cores we can get the solution time down to about 1 day, with 26 cores we can get it down to about a half days, etc. If we assume the computation time is directly proportional to the size of the data (but often it scales nonlinearly), then using 500,000 records instead of 5 million records would reduce our total computation time from hours for a single core down to 31.5 hours. This is still a significant period of time, but much more acceptable when parallelised, particularly when
6 we still have sufficient statistical power to identify patterns that will have real business value. But, we forgot one very important step in the model building process - we still need to search for the best combination of predictor variables, and every separate combination of those will require another 3750 separate model fittings! This example is deliberately simplistic. There are sophisticated methods for reducing the total number of times a model needs be fitted, but the important point is that the total number of fittings required to build a production-ready model is usually much higher than many people realise. And the more data we use, the longer the model development process will take. 4. Model testing the path to true enlightenment: This is, in my opinion, the single most important phase of predictive modelling. It is when we try to simulate a production setting by using the model to make predictions based on hold-out data that the model has never seen before. These predictions are then compared to the actual historical outcomes from the hold-out data and it is this comparison, and this comparison only, that we use to evaluate the effectiveness of our model. But there's a catch the technically correct way to use hold-out data is to only ever use it once. Say we build a model that did not perform well on our hold-out data. We start again, but to test the new model, we need a new set of hold-out data. If we don't have new hold-out data, all we do is enter a process of optimising our models for a single set of hold-out data. Consequently, model testing is where big data, or at least lots of data, really can be beneficial. In an ideal scenario, we would want to test our models many times using independent sets of hold-out data to get a good understanding of model performance before we risk putting them into production. But we are still not finished 5. Model maintenance: Heraclitus (535 BC 475 BC) is attributed with saying "There is nothing permanent except change". What was true then remains true today we must expect any system we model to change over time, possibly to the extent that our models begin to fail. This is particularly relevant in situations where we are using models to manage or control a system. In this case, the actions of management and control ARE changing the system, so we need to be particularly vigilant of the performance of such models over time. In the best case scenario, our models may simply need to be retrained or retuned on a regular basis to accommodate change. But the system may also change to such a degree that our original modelling paradigm may no longer be relevant and we are forced to go through the entire modelling process again. A classic example of a model failing over time is the Google flu predictor. The folks at Google are clever, and I m confident their model was developed using sound statistical principles. But its release to the public apparently influenced the incidence of the term "flu" appearing on the internet to such an extent that it caused the model to seriously overestimate the next flu outbreak. 6. Integration and model deployment: As mentioned previously, it doesn't matter how good a predictive model is if it remains inaccessible within an organisation. In order for predictive analytics to provide the greatest return on investment, it is essential that people within the organisation feel empowered in their roles through using the insight the models provide. When people feel empowered in this manner, the culture of the organisation will begin to embrace the use of predictive models as a valuable tool that helps people do their jobs better. This type of empowerment comes when the output from the models is reliable, easily accessible, easily interpreted, and meaningful in the context of people's daily roles and responsibilities.
7 One of the great benefits of the recent popularisation of big data and analytics is the wealth of products available today that make the integration and deployment of predictive analytics easier than it ever has been. All the leading commercial vendors in the ERP, data warehousing, business intelligence, and analytics markets have frameworks for deploying models and presenting the results to business users typically through some form of web interface. Many of these vendors also have cloud-based services that are targeted at small and medium businesses, providing a compelling basis for smaller organisations to participate in the benefits that well-designed predictive modelling can provide. Nevertheless, the choice of software for building and deploying models is an important part of any business case for predictive analytics. At a coarse level, there is a choice between commercial products and open source products. Fundamentally (though not strictly true in all cases), at the level of computer code operating on data and returning numerical results, it can be argued that there is not a lot to distinguish the commercial products from the open source products they all perform pretty much as expected. One of the real differences though lies in the level of technical and product support that is available. Generally speaking, help will usually be available quickly if something goes wrong with a commercial product. This is not necessarily true, nor false, for open source products. But an organisation that chooses to use open source products may need to expect to engage technical experts in those products at a frequency that may be higher than for a commercial product. Contractors and consultants who specialise in open source products continue to grow in number, so the open source pathway need not be constraining or unwise. It is up to the individual organisation to determine what it values most (for example, service versus cost) and the longer-term costs and benefits that may be associated with that preference (for example, bespoke solutions based on open source products and built by consultants or contractors may not provide adequate long-term flexibility or upgradeability). Again, the more aware an organisation can be regarding these alternatives, the better it will be placed to make good technical and business choices. Summary There are a number of important issues that organisations need to consider carefully before embarking on a programme of predictive analytics. Some of those issues have been touched on here, but many have not. And while these issues may seem daunting, the intent of this discussion has not been to dissuade people from investigating the utility of predictive analytics for their organisation. Instead, the intent has been to assist those people by exposing at least some of the issues they will need to consider. Predictive analytics is likely to become a necessary function for many organisations. In the public sectors such as healthcare, this will be driven by continued reductions in funding and increases in public and regulatory expectations. In the private sector, increasing competition will drive many businesses in this direction. Even though the benefits that an organisation may realise through the careful and considered application of predictive analytics could be substantial, it is worth keeping in mind that for every publicised success story, there is likely to be numerous expensive and/or embarrassing failures which we simply do not hear about. It is this type of failure that I hope the following five take-home summary points will help minimise: 1. Understand your business requirements and develop a robust business case for implementing predictive analytics. 2. Get the right people involved early in the journey. 3. Get the right technology for your organisation's requirements. Be rational and pragmatic but make sure your choices are capable of meeting your projected growth. Technology choices also need to
8 take into account the availability of suitably skilled staff or contractors in addition to licensing and maintenance fees, etc. 4. Always subject predictive models to rigorous testing using hold-out data that was not used during any stage of model development. This is the only way to understand if a model can truly perform in a production environment. 5. Ensure the model output is reliable, easy to access, easy to interpret, and that it helps empower people in their daily roles.
Prescriptive Analytics. A business guide
Prescriptive Analytics A business guide May 2014 Contents 3 The Business Value of Prescriptive Analytics 4 What is Prescriptive Analytics? 6 Prescriptive Analytics Methods 7 Integration 8 Business Applications
10 things you should look for. Choosing HR software
10 things you should look for Choosing HR software Introduction Selecting a new piece of HR software can be a daunting task. There s a lot to think about. At the end of the day, the chosen software won
Measuring and Monitoring the Quality of Master Data By Thomas Ravn and Martin Høedholt, November 2008
Measuring and Monitoring the Quality of Master Data By Thomas Ravn and Martin Høedholt, November 2008 Introduction We ve all heard about the importance of data quality in our IT-systems and how the data
Elevator Service Preventive or Predictive
www.wipro.com Elevator Service Preventive or Predictive Market Differentiation Through Remote Monitoring Data and a Predictive Service Program Russell Gray Sr. Consultant, Business Process Group Aftermarket
Media attribution: Optimising digital marketing spend in Financial Services
Media : Optimising digital marketing spend in Financial Services A multi touch study on the role of Facebook, display and search marketing in Financial Services. Executive summary Media solves the key
The reality of cloud. Go beyond the hype and make a better choice. t 0845 5055 365 e [email protected]. www.365itms.co.uk
The reality of cloud Go beyond the hype and make a better choice www. The meaning of cloud 1. Cloud means different things to different people, something that s reflected in the many definitions of what
Correlational Research
Correlational Research Chapter Fifteen Correlational Research Chapter Fifteen Bring folder of readings The Nature of Correlational Research Correlational Research is also known as Associational Research.
Marketing Mix Modelling and Big Data P. M Cain
1) Introduction Marketing Mix Modelling and Big Data P. M Cain Big data is generally defined in terms of the volume and variety of structured and unstructured information. Whereas structured data is stored
10 Golden Rules of Project Risk Management
10 Golden Rules of Project Risk Management By Bart Jutte The benefits of risk management in projects are huge. You can gain a lot of money if you deal with uncertain project events in a proactive manner.
The Role of Feedback Management in Becoming Customer Centric
A Customer Centricity White Paper 5 Old Coach Road Hudson, NH 03051 603-491-7948 www.customercentricity.biz The Role of Feedback Management in Becoming Customer Centric By Stewart Nash QuestBack Boston
Conversion Rate Optimisation Guide
Conversion Rate Optimisation Guide Improve the lead generation performance of your website - Conversion Rate Optimisation in a B2B environment Why read this guide? Work out how much revenue CRO could increase
Basel Committee on Banking Supervision. Working Paper No. 17
Basel Committee on Banking Supervision Working Paper No. 17 Vendor models for credit risk measurement and management Observations from a review of selected models February 2010 The Working Papers of the
Navigating Big Data business analytics
mwd a d v i s o r s Navigating Big Data business analytics Helena Schwenk A special report prepared for Actuate May 2013 This report is the third in a series and focuses principally on explaining what
Deploying the Enterprise Cloud
Deploying the Enterprise Cloud Tim Hancock Publication date: October 2012 Deploying the Enterprise Cloud Many enterprises now recognise that the benefits outweigh the risks in Cloud adoption. However,
Criminal Justice Evaluation Framework (CJEF): Conducting effective outcome evaluations
Criminal Justice Research Department of Premier and Cabinet Criminal Justice Evaluation Framework (CJEF): Conducting effective outcome evaluations THE CRIMINAL JUSTICE EVALUATION FRAMEWORK (CJEF) The Criminal
Digital Strategy. Digital Strategy. 2015 CGI IT UK Ltd. Digital Innovation. Enablement Services
Digital Strategy Digital Strategy Digital Innovation Enablement Services 2015 CGI IT UK Ltd. Contents Digital strategy overview Business drivers Anatomy of a solution Digital strategy in practice Delivery
Organizing Your Approach to a Data Analysis
Biost/Stat 578 B: Data Analysis Emerson, September 29, 2003 Handout #1 Organizing Your Approach to a Data Analysis The general theme should be to maximize thinking about the data analysis and to minimize
Agile Project Portfolio Management
C an Do GmbH Implerstr. 36 81371 Munich, Germany AGILE PROJECT PORTFOLIO MANAGEMENT Virtual business worlds with modern simulation software In view of the ever increasing, rapid changes of conditions,
Conn Valuation Services Ltd.
CAPITALIZED EARNINGS VS. DISCOUNTED CASH FLOW: Which is the more accurate business valuation tool? By Richard R. Conn CMA, MBA, CPA, ABV, ERP Is the capitalized earnings 1 method or discounted cash flow
ADDING NETWORK INTELLIGENCE TO VULNERABILITY MANAGEMENT
ADDING NETWORK INTELLIGENCE INTRODUCTION Vulnerability management is crucial to network security. Not only are known vulnerabilities propagating dramatically, but so is their severity and complexity. Organizations
THE THREE "Rs" OF PREDICTIVE ANALYTICS
THE THREE "Rs" OF PREDICTIVE As companies commit to big data and data-driven decision making, the demand for predictive analytics has never been greater. While each day seems to bring another story of
White Paper: Impact of Inventory on Network Design
White Paper: Impact of Inventory on Network Design Written as Research Project at Georgia Tech with support from people at IBM, Northwestern, and Opex Analytics Released: January 2014 Georgia Tech Team
THE PREDICTIVE MODELLING PROCESS
THE PREDICTIVE MODELLING PROCESS Models are used extensively in business and have an important role to play in sound decision making. This paper is intended for people who need to understand the process
The metrics that matter
WHITE PAPER The metrics that matter How actionable analytics can transform field service management performance. www. Introduction The top strategic action for two-thirds of service organisations is to
Analytics in Days White Paper and Business Case
Analytics in Days White Paper and Business Case Analytics Navigating the Maze Analytics is hot. It seems virtually everyone needs or wants it, but many still aren t sure what the business case is or how
Data Mining Applications in Higher Education
Executive report Data Mining Applications in Higher Education Jing Luan, PhD Chief Planning and Research Officer, Cabrillo College Founder, Knowledge Discovery Laboratories Table of contents Introduction..............................................................2
Achieving the Gold Standard in Paid-Search Bid Optimization
MORE MATH. BETTER MATH. BEST ROI. Achieving the Gold Standard in Paid-Search Bid Optimization Understanding the benefits of global keyword-level modeling white paper: The Gold Standard [email protected]
KEEPING CUSTOMERS USING ANALYTICS
KEEPING CUSTOMERS USING ANALYTICS This paper outlines a robust approach to investigating and managing customer churn for those in the business-to-consumer market. In order to address customer retention
Developing In-House Vs. Off the Shelf. - A white paper by Clydebuilt Business Solutions Ltd
Developing In-House Vs. Off the Shelf - A white paper by Clydebuilt Business Solutions Ltd Developing In House vs. Off the Shelf Naturally, as a software development company that operates solely within
Technology Integration and the Service Desk
Technology Integration and the Service Desk Table of Contents Introduction 1 Key Findings 2 What benefits do you associate with the integration of technologies? 3 When considering your service management
What? So what? NOW WHAT? Presenting metrics to get results
What? So what? NOW WHAT? What? So what? Visualization is like photography. Impact is a function of focus, illumination, and perspective. What? NOW WHAT? Don t Launch! Prevent your own disastrous decisions
ElegantJ BI. White Paper. The Competitive Advantage of Business Intelligence (BI) Forecasting and Predictive Analysis
ElegantJ BI White Paper The Competitive Advantage of Business Intelligence (BI) Forecasting and Predictive Analysis Integrated Business Intelligence and Reporting for Performance Management, Operational
DON T GET LOST IN THE FOG OF BIG DATA
DON T GET LOST IN THE FOG OF BIG DATA MERCER S LESSONS FOR SUCCESS IN WORKFORCE ANALYTICS If 2013 has produced a breakthrough technology phrase, it is big data, a fairly vague but forceful term that features
Behavioral Segmentation
Behavioral Segmentation TM Contents 1. The Importance of Segmentation in Contemporary Marketing... 2 2. Traditional Methods of Segmentation and their Limitations... 2 2.1 Lack of Homogeneity... 3 2.2 Determining
Data Center Consolidation
Data Center Consolidation With computers spread across the enterprise, and the number of networked databases and applications growing at an exponential rate, the challenges of efficient networking, data
An effective approach to preventing application fraud. Experian Fraud Analytics
An effective approach to preventing application fraud Experian Fraud Analytics The growing threat of application fraud Fraud attacks are increasing across the world Application fraud is a rapidly growing
Advanced Big Data Analytics with R and Hadoop
REVOLUTION ANALYTICS WHITE PAPER Advanced Big Data Analytics with R and Hadoop 'Big Data' Analytics as a Competitive Advantage Big Analytics delivers competitive advantage in two ways compared to the traditional
W H I T E P A P E R. Deriving Intelligence from Large Data Using Hadoop and Applying Analytics. Abstract
W H I T E P A P E R Deriving Intelligence from Large Data Using Hadoop and Applying Analytics Abstract This white paper is focused on discussing the challenges facing large scale data processing and the
WHITE PAPER AUTOMATED, REAL-TIME RISK ANALYSIS AND REMEDIATION
WHITE PAPER AUTOMATED, REAL-TIME RISK ANALYSIS AND REMEDIATION Table of Contents Executive Summary...3 Vulnerability Scanners Alone Are Not Enough...3 Real-Time Change Configuration Notification is the
Information Visualization WS 2013/14 11 Visual Analytics
1 11.1 Definitions and Motivation Lot of research and papers in this emerging field: Visual Analytics: Scope and Challenges of Keim et al. Illuminating the path of Thomas and Cook 2 11.1 Definitions and
IBM Cloud Managed Infrastructure Services for New Zealand Government
IBM Cloud Managed Infrastructure Services for New Zealand Government An agile cloud infrastructure designed to provide quick access to a security- rich, enterprise- class virtual server environment. 2
ANALYTICS A FUTURE IN ANALYTICS
ANALYTICS A FUTURE IN ANALYTICS WHAT IS ANALYTICS? In the information age in which we live, almost all of us consume and produce digital data, either for business, community or private uses. We access
How to Plan a Successful Load Testing Programme for today s websites
How to Plan a Successful Load Testing Programme for today s websites This guide introduces best practise for load testing to overcome the complexities of today s rich, dynamic websites. It includes 10
Frequency Matters. The keys to optimizing email send frequency
The keys to optimizing email send frequency Email send frequency requires a delicate balance. Send too little and you miss out on sales opportunities and end up leaving money on the table. Send too much
CPNI VIEWPOINT CONFIGURING AND MANAGING REMOTE ACCESS FOR INDUSTRIAL CONTROL SYSTEMS
CPNI VIEWPOINT CONFIGURING AND MANAGING REMOTE ACCESS FOR INDUSTRIAL CONTROL SYSTEMS MARCH 2011 Acknowledgements This Viewpoint is based upon the Recommended Practice: Configuring and Managing Remote Access
Testing Rails. by Josh Steiner. thoughtbot
Testing Rails by Josh Steiner thoughtbot Testing Rails Josh Steiner April 10, 2015 Contents thoughtbot Books iii Contact us................................ iii Introduction 1 Why test?.................................
Maximizing the ROI Of Visual Rules
Table of Contents Introduction... 3 Decision Management... 3 Decision Discovery... 4 Decision Services... 6 Decision Analysis... 11 Conclusion... 12 About Decision Management Solutions... 12 Acknowledgements
SEVEN WAYS TO AVOID ERP IMPLEMENTATION FAILURE SPECIAL REPORT SERIES ERP IN 2014 AND BEYOND
SEVEN WAYS TO AVOID ERP IMPLEMENTATION FAILURE SPECIAL REPORT SERIES ERP IN 2014 AND BEYOND CONTENTS INTRODUCTION................................ 3 UNDERSTAND YOUR CURRENT SITUATION............ 4 DRAW
Instilling Confidence in Security and Risk Operations with Behavioral Analytics and Contextualization
WHITEPAPER Instilling Confidence in Security and Risk Operations with Behavioral Analytics and Contextualization Understanding Why Automated Machine Learning Behavioral Analytics with Contextualization
BPM and Simulation. A White Paper. Signavio, Inc. Nov 2013. Katharina Clauberg, William Thomas
BPM and Simulation A White Paper Signavio, Inc. Nov 2013 Katharina Clauberg, William Thomas Table of Contents 1. Executive Summary... 3 2. Setting the Scene for Process Change... 4 3. Identifying the Goals
BEST PRACTICES IN DEMAND AND INVENTORY PLANNING
WHITEPAPER BEST PRACTICES IN DEMAND AND INVENTORY PLANNING for Food & Beverage Companies WHITEPAPER BEST PRACTICES IN DEMAND AND INVENTORY PLANNING 2 ABOUT In support of its present and future customers,
IS YOUR DATA WAREHOUSE SUCCESSFUL? Developing a Data Warehouse Process that responds to the needs of the Enterprise.
IS YOUR DATA WAREHOUSE SUCCESSFUL? Developing a Data Warehouse Process that responds to the needs of the Enterprise. Peter R. Welbrock Smith-Hanley Consulting Group Philadelphia, PA ABSTRACT Developing
SEVEN WAYS THAT BUSINESS PROCESS MANAGEMENT CAN IMPROVE YOUR ERP IMPLEMENTATION SPECIAL REPORT SERIES ERP IN 2014 AND BEYOND
SEVEN WAYS THAT BUSINESS PROCESS MANAGEMENT CAN IMPROVE YOUR ERP IMPLEMENTATION SPECIAL REPORT SERIES ERP IN 2014 AND BEYOND CONTENTS INTRODUCTION 3 EFFECTIVELY MANAGE THE SCOPE OF YOUR IMPLEMENTATION
Survey of more than 1,500 Auditors Concludes that Audit Professionals are Not Maximizing Use of Available Audit Technology
Survey of more than 1,500 Auditors Concludes that Audit Professionals are Not Maximizing Use of Available Audit Technology Key findings from the survey include: while audit software tools have been available
EFFECTIVE VENDOR MANAGEMENT: REAPING LONG-TERM BENEFITS FROM YOUR VENDOR RELATIONSHIPS
EFFECTIVE VENDOR MANAGEMENT: REAPING LONG-TERM BENEFITS FROM YOUR VENDOR RELATIONSHIPS TERRA FIRMA, AUGUST 2013 Leading organisations have understood for some time that active vendor management, as opposed
Introduction to quantitative research
8725 AR.qxd 25/08/2010 16:36 Page 1 1 Introduction to quantitative research 1.1. What is quantitative research? Research methods in education (and the other social sciences) are often divided into two
Certificate Program in Applied Big Data Analytics in Dubai. A Collaborative Program offered by INSOFE and Synergy-BI
Certificate Program in Applied Big Data Analytics in Dubai A Collaborative Program offered by INSOFE and Synergy-BI Program Overview Today s manager needs to be extremely data savvy. They need to work
Risk Analysis Overview
What Is Risk? Uncertainty about a situation can often indicate risk, which is the possibility of loss, damage, or any other undesirable event. Most people desire low risk, which would translate to a high
PREDICTIVE ANALYTICS vs HOT SPOTTING
PREDICTIVE ANALYTICS vs HOT SPOTTING A STUDY OF CRIME PREVENTION ACCURACY AND EFFICIENCY 2014 EXECUTIVE SUMMARY For the last 20 years, Hot Spots have become law enforcement s predominant tool for crime
Software Metrics: Roadmap
Software Metrics: Roadmap By Norman E. Fenton and Martin Neil Presentation by Karim Dhambri Authors (1/2) Norman Fenton is Professor of Computing at Queen Mary (University of London) and is also Chief
Master Data Management
ADVISORY Master Data Management It s not just about a new concept, it s about bringing real value to the business Data is at the heart of nearly any business. Being able to rely on the consistency and
OPTIMUS SBR. Optimizing Results with Business Intelligence Governance CHOICE TOOLS. PRECISION AIM. BOLD ATTITUDE.
OPTIMUS SBR CHOICE TOOLS. PRECISION AIM. BOLD ATTITUDE. Optimizing Results with Business Intelligence Governance This paper investigates the importance of establishing a robust Business Intelligence (BI)
8. KNOWLEDGE BASED SYSTEMS IN MANUFACTURING SIMULATION
- 1-8. KNOWLEDGE BASED SYSTEMS IN MANUFACTURING SIMULATION 8.1 Introduction 8.1.1 Summary introduction The first part of this section gives a brief overview of some of the different uses of expert systems
Automation can dramatically increase product quality, leading to lower field service, product support and
QA Automation for Testing Medical Device Software Benefits, Myths and Requirements Automation can dramatically increase product quality, leading to lower field service, product support and liability cost.
White Paper: Migrating Email to the Cloud
White Paper: Migrating Email to the Cloud 2015, Cloud Point ltd. All rights reserved. INTELLECTUAL PROPERTY DISCLAIMER This white paper is for informational purposes only and is provided as is with no
Better decision making under uncertain conditions using Monte Carlo Simulation
IBM Software Business Analytics IBM SPSS Statistics Better decision making under uncertain conditions using Monte Carlo Simulation Monte Carlo simulation and risk analysis techniques in IBM SPSS Statistics
BENCHMARKING PERFORMANCE AND EFFICIENCY OF YOUR BILLING PROCESS WHERE TO BEGIN
BENCHMARKING PERFORMANCE AND EFFICIENCY OF YOUR BILLING PROCESS WHERE TO BEGIN There have been few if any meaningful benchmark analyses available for revenue cycle management performance. Today that has
The Challenge of Helping Adults Learn: Principles for Teaching Technical Information to Adults
The Challenge of Helping Adults Learn: Principles for Teaching Technical Information to Adults S. Joseph Levine, Ph.D. Michigan State University [email protected] One of a series of workshop handouts made
In this presentation, you will be introduced to data mining and the relationship with meaningful use.
In this presentation, you will be introduced to data mining and the relationship with meaningful use. Data mining refers to the art and science of intelligent data analysis. It is the application of machine
Compliance. Technology. Process. Using Automated Decisioning and Business Rules to Improve Real-time Risk Management
Technology Process Compliance Using Automated Decisioning and Business Rules to Improve Real-time Risk Management Sandeep Gupta, Equifax James Taylor, Smart (enough) Systems August 2008 Equifax is a registered
Using Adaptive Web Design to Increase E-Commerce Sales & Lead Generation
Using Adaptive Web Design to Increase E-Commerce Sales & Lead Generation Building High Conversion Websites Copyright 2012-2014 Published by Click Laboratory, LLC 4208 Sandy Shores, Lutz, Florida 33558
White Paper. Cloud Computing. Effective Web Solution Technology Investment. January 2011. www.k-bc.co.uk
White Paper Cloud Computing Effective Web Solution Technology Investment January 2011 www.k-bc.co.uk Cloud Computing The future of effective technology investment The benefit of web solution technology
1.877.896.3611 www.infrascale.com
White Paper Making the Upgrade: From Backup Vendor to Data Protection Platform nfrascaletm Simplify Your Business and Offer Customers Complete Data Protection Your customers are ready to explore data protection
A GUIDE TO IMPLEMENTING SAP BUSINESS ONE
A GUIDE TO IMPLEMENTING SAP BUSINESS ONE THE ULTIMATE GUIDE If you are considering SAP Business One as your business management solution or have already selected SAP Business One and you are about to embark
Business Intelligence 100 Success Secrets
Business Intelligence 100 Success Secrets 100 Most Asked Questions: The Missing BI software, Tools, Consulting and Solutions Guide Gerard Blokdijk Business Intelligence 100 Success Secrets Copyright 2008
How to Write a Successful PhD Dissertation Proposal
How to Write a Successful PhD Dissertation Proposal Before considering the "how", we should probably spend a few minutes on the "why." The obvious things certainly apply; i.e.: 1. to develop a roadmap
Aligning CFO and CIO Priorities
whitepaper economics Aligning and Priorities Forward-thinking organizations are viewing computing as an investment in business transformation, not just a way to cut costs for IT. Thanks to the, s and s
HR STILL GETTING IT WRONG BIG DATA & PREDICTIVE ANALYTICS THE RIGHT WAY
HR STILL GETTING IT WRONG BIG DATA & PREDICTIVE ANALYTICS THE RIGHT WAY OVERVIEW Research cited by Forbes estimates that more than half of companies sampled (over 60%) are investing in big data and predictive
Gerard Mc Nulty Systems Optimisation Ltd [email protected]/0876697867 BA.,B.A.I.,C.Eng.,F.I.E.I
Gerard Mc Nulty Systems Optimisation Ltd [email protected]/0876697867 BA.,B.A.I.,C.Eng.,F.I.E.I Data is Important because it: Helps in Corporate Aims Basis of Business Decisions Engineering Decisions Energy
7 things to ask when upgrading your ERP solution
Industrial Manufacturing 7 things to ask when upgrading your ERP solution The capabilities gap between older versions of ERP designs and current designs can create a problem that many organizations are
What Is Specific in Load Testing?
What Is Specific in Load Testing? Testing of multi-user applications under realistic and stress loads is really the only way to ensure appropriate performance and reliability in production. Load testing
TRUSTING YOUR DATA. How companies benefit from a reliable data using Information Steward 1/2/2013
TRUSTING YOUR DATA 1/2/2013 How companies benefit from a reliable data using Information Steward How successful is your company s data migration and conversion? How do you quantify and measure the reliability
