Using Adaptive Random Trees (ART) for optimal scorecard segmentation



Similar documents
Summary. WHITE PAPER Using Segmented Models for Better Decisions

KnowledgeSTUDIO HIGH-PERFORMANCE PREDICTIVE ANALYTICS USING ADVANCED MODELING TECHNIQUES

Customer Management TRIAD version 2.0

ElegantJ BI. White Paper. The Competitive Advantage of Business Intelligence (BI) Forecasting and Predictive Analysis

Debt Manager solution overview

KnowledgeSEEKER POWERFUL SEGMENTATION, STRATEGY DESIGN AND VISUALIZATION SOFTWARE

Easily Identify Your Best Customers

Lecture 10: Regression Trees

A Basic Guide to Modeling Techniques for All Direct Marketing Challenges

Understanding the Benefits of IBM SPSS Statistics Server

Data Mining Applications in Higher Education

WHITEPAPER. How to Credit Score with Predictive Analytics

KnowledgeSEEKER Marketing Edition

Product recommendations and promotions (couponing and discounts) Cross-sell and Upsell strategies

Data-Driven Decisions: Role of Operations Research in Business Analytics

A Property & Casualty Insurance Predictive Modeling Process in SAS

Get to Know the IBM SPSS Product Portfolio

How to Optimize Your Data Mining Environment

Nine Common Types of Data Mining Techniques Used in Predictive Analytics

Benchmarking of different classes of models used for credit scoring

NICE SALES PERFORMANCE MANAGEMENT (SPM)

Data Mining Algorithms Part 1. Dejan Sarka

WHITEPAPER. Creating and Deploying Predictive Strategies that Drive Customer Value in Marketing, Sales and Risk

WebFOCUS RStat. RStat. Predict the Future and Make Effective Decisions Today. WebFOCUS RStat

The Predictive Data Mining Revolution in Scorecards:

SQL Server 2005 Features Comparison

Gerry Hobbs, Department of Statistics, West Virginia University

Five Predictive Imperatives for Maximizing Customer Value

Today, the world s leading insurers

IBM's Fraud and Abuse, Analytics and Management Solution

Business Intelligence and Big Data Analytics: Speeding the Cycle from Insights to Action Four Steps to More Profitable Customer Engagement

Successfully Implementing Predictive Analytics in Direct Marketing

EXPLORING & MODELING USING INTERACTIVE DECISION TREES IN SAS ENTERPRISE MINER. Copyr i g ht 2013, SAS Ins titut e Inc. All rights res er ve d.

TMO (The Mortgage Originator) product overview

How To Make A Credit Risk Model For A Bank Account

Lakeside. Lakeside Software and IBM: Statement of Fact. SysTrack. A Lakeside Software White Paper November 2011

In-Database Analytics

An Introduction to Data Mining. Big Data World. Related Fields and Disciplines. What is Data Mining? 2/12/2015

Decision Trees from large Databases: SLIQ

Advanced Case Management. Chris den Hoedt

Five Predictive Imperatives for Maximizing Customer Value

Lawson Talent Management

Predictive Modeling and Big Data

Patient Relationship Management

WHITE PAPER. Harnessing the Power of Advanced Analytics How an appliance approach simplifies the use of advanced analytics

Five predictive imperatives for maximizing customer value

HOW PREDICTIVE ANALYTICS DRIVES PROFITABILITY IN ASSET FINANCE

Data Mining Practical Machine Learning Tools and Techniques

Decision Trees What Are They?

BIDM Project. Predicting the contract type for IT/ITES outsourcing contracts

ElegantJ BI. White Paper. Achieve a Complete Business Picture with a Business Intelligence (BI) Dashboard

Why Ensembles Win Data Mining Competitions

Predictive Analytics for Donor Management

Variable Selection in the Credit Card Industry Moez Hababou, Alec Y. Cheng, and Ray Falk, Royal Bank of Scotland, Bridgeport, CT

A Robust Method for Solving Transcendental Equations

THE HYBRID CART-LOGIT MODEL IN CLASSIFICATION AND DATA MINING. Dan Steinberg and N. Scott Cardell

Loss Forecasting Methodologies Approaches to Successful Risk Management

Compliance. Technology. Process. Using Automated Decisioning and Business Rules to Improve Real-time Risk Management

Comply and Compete: Model Management Best Practices

Aspen Collaborative Demand Manager

IBM Global Business Services Microsoft Dynamics CRM solutions from IBM

A fast, powerful data mining workbench designed for small to midsize organizations

Data Mining is the process of knowledge discovery involving finding

Oracle Real Time Decisions

ICD-10 Advantages Require Advanced Analytics

How the Past Changes the Future of Fraud

Necto on Azure The Ultimate Cloud Solution for BI

DATA MINING TECHNIQUES AND APPLICATIONS

Predictive Analytics Software Suite

Improving the Performance of Data Mining Models with Data Preparation Using SAS Enterprise Miner Ricardo Galante, SAS Institute Brasil, São Paulo, SP

Data Mining - Evaluation of Classifiers

STATISTICA Solutions for Financial Risk Management Management and Validated Compliance Solutions for the Banking Industry (Basel II)

Business Information Services. Product overview

Beating the MLB Moneyline

BPM for Structural Integrity Management in Oil and Gas Industry

Data Mining Classification: Decision Trees

Understanding Characteristics of Caravan Insurance Policy Buyer

How To Be Successful In The Czech Republic

DISCOVER MERCHANT PREDICTOR MODEL

Unlock the business value of enterprise data with in-database analytics

Pentaho Data Mining Last Modified on January 22, 2007

Learning Example. Machine learning and our focus. Another Example. An example: data (loan application) The data and the goal

Predictive Analytics: Achieving Greater Decision Accuracy, Better Risk Segmentation, and Greater Profitability

College Readiness LINKING STUDY

Prediction of Stock Performance Using Analytical Techniques

Business Intelligence. Tutorial for Rapid Miner (Advanced Decision Tree and CRISP-DM Model with an example of Market Segmentation*)

Building a Question Classifier for a TREC-Style Question Answering System

Principles of Data Mining by Hand&Mannila&Smyth

Telecommunications, Media, and Technology. Leveraging big data to optimize digital marketing

Alex Vidras, David Tysinger. Merkle Inc.

Full and Complete Binary Trees

analytics stone Automated Analytics and Predictive Modeling A White Paper by Stone Analytics

4 WHAT IS PREDICTIVE ANALYTICS? 6 WHO USES PREDICTIVE ANALYTICS AND HOW? 10 WHAT ARE THE TYPES OF PREDICTIVE ANALYTICS?

Three Ways to Improve Claims Management with Business Analytics

Paper AA Get the highest bangs for your marketing bucks using Incremental Response Models in SAS Enterprise Miner TM

Transcription:

A FAIR ISAAC WHITE PAPER Using Adaptive Random Trees (ART) for optimal scorecard segmentation By Chris Ralph Analytic Science Director April 2006 Summary Segmented systems of models are widely recognized as a highly effective way of boosting predictive power; however, the traditional process for building these systems is often considered too laborious to justify the effort. Fair Isaac uses Adaptive Random Trees (ART) technology to create segmented models that predict outcomes with great precision. ART uses a powerful global procedure that looks at the entire terrain before determining the best segmentation scheme or tree, rather than local branch by branch decisions, as in the traditional manual techniques. Rooted in advances in genetic algorithms, ART automates the process of searching thousands of combinations of segmentations and models to find the best segmented model system. FOR MORE phone from US worldwide email web INFORMATION 1 800 999 2955 1 612 758 5200 info@fairisaac.com www.fairisaac.com

Table of Contents Segmented Scorecard Development...1 Traditional approach to segmented scorecard development... 2 ART approach to segmented scorecard development... 2 How ART works...3 ART in practice... 5 ART availability... 5 About Fair Isaac...6 2006 Fair Isaac Corporation. All rights reserved. i

Segmented Scorecard Development The process of predictive model development requires the identification of features in the data that exhibit clear and interpretable patterns with respect to the target variable of interest. Consider, for example, an analyst charged with the task of ranking a population of prospective customers by the likelihood that each will respond to an offer for a new low interest credit card. One of the key predictors of responsiveness to this offer is age, which exhibits the following pattern: In other words, when the entire population is ranked by age, the analysts notices that younger prospects are more likely to respond, and older prospects are less likely to respond. This discovery could be leveraged to accomplish the task at hand by including age in a predictive model, or scorecard, such that younger prospects get a higher score (i.e., indicating higher response likelihood) and older prospects get a lower score. To build a good scorecard, the analyst needs to continue searching through the data to find other predictors, in addition to age, that lead to other useful patterns with respect to the target that can also be leveraged. During the course of this investigation, however, the analyst also notices that not all younger prospects are highly likely to respond to this particular offer. In fact, the nice, intuitive response pattern shown above begins to look quite different when the population is split into sub-segments representing different levels of income. 2006 Fair Isaac Corporation. All rights reserved. 1

By splitting the population into sub-segments by income, the analyst is able to identify an important fact. It turns out that the relation of response to age is much stronger in the low income sub-segment when compared to the overall population. In fact, it appears that age becomes almost useless as a predictor of response when income is high. Rather than build a single scorecard for the entire population, this discovery leads the analyst to build a separate scorecard in each of these sub-populations created by splitting on income, with age included as a predictor in the low income scorecard but not in the high income scorecard. As the analyst continues with her analysis in these two sub-populations, she discovers some additional variables that work well in the high-income group that are not predictive in the low-income group, and vice-versa. This leads her to develop a segmented system of models that generates much more accurate estimates of responsiveness than can be accomplished with a single, global model. Traditional approach to segmented scorecard development The scenario outlined above represents a very common situation faced by analysts who are trying to build the best possible predictions across a wide range of modeling problems. Segmented systems of models are widely recognized as a highly effective way of boosting predictive power, however the traditional process to build these systems is often considered much too laborious to justify the effort. In the absence of an automated approach, most analysts use a manual or greedy search process. Typically, this involves starting with the entire population and manually selecting segmentation variables. Such as income in the example above; splitting thresholds within those variables one at a time; and building scorecards in each resulting sub-population. The winning split becomes the one that results in the greatest overall boost in predictive power. The process is then repeated in each of the chosen new sub-populations until either predictive power stops improving, or the analysts run out of time or energy. Fair Isaac has been a worldwide leader in the development of highly effective and robust segmented scorecard systems for more than 20 years. This is due largely to our ability to develop unique technologies that help to streamline the search process.and identify the best combination of splitters and split points to define an overall segmentation tree. Our scorecard technology will construct robust predictive models in the leaf nodes of those trees. However, until recently, these powerful segmentation tools still suffered from the pitfalls of a greedy search process. The term greedy stems from the fact that each splitting decision is made using only information from the sub-population that is currently being split. Although we had been able to develop procedures and tools that ensured the selected segments led to robust and effective solutions, this traditional process was inherently near-sighted, because it was incapable of calculating the effects of local decisions i.e., Those regarding the current sub-population on downstream segments. In other words, a splitting decision made near the top of a decision tree could lead to subsequent sub-optimal splits that adversely affect the overall system of scorecards. ART approach to segmented scorecard development Adaptive Random Trees is a highly innovative technology that not only addresses the greedy search problem, but also provides an intelligent, automated process to find powerful segmentation systems in a small fraction of the time required for the manual approach described above. ART is designed to work with binary target variables. With a little bit of clever data pre-processing, ART can also be used on continuous outcome targets. Rather than search through sets of splitters and split points in successive order to build a segmentation tree one level at a time, ART works by evolving an entire population of fully defined trees over time towards increasing overall predictive power. Controlling this search is a proprietary genetic algorithm, which works by selecting sets of the most effective trees and swapping information i.e., splitters and split points between them to create successive generations of new trees. 2006 Fair Isaac Corporation. All rights reserved. 2

Over time, as this process continues, the best overall combination of splitters and split points emerges from this population, and the analyst is provided with the segmentation logic for a system that has been optimized with a global objective function that measures overall predictive power. How ART works The first step in developing a segmented scorecard system using ART is to develop an initial parent scorecard on the entire population. Not only does this provide a benchmark against which the segmented systems can be compared, it is also used to define a set of binned predictor (independent) variables that will appear in the segment-level models in every leaf node of each tree that ART considers. Instead of calculating a single coefficient for each predictor, as in a linear regression model, the scorecards work by defining a set of carefully chosen bins within each predictor variable and associating a score weight to each of those bins. This approach enables the resulting scorecard to take full advantage of the underlying patterns inside each predictor with respect to the target, including any non-linear patterns that are often missed by traditional regression techniques. Although every leaf-node scorecard uses a common set of binned predictors, the resulting models are specifically predictive for that sub-population because they contain segment-specific bin-level score weights. In ART, those weights are set to the weight of evidence values calculated within the bins. WoE = log i e P[ bini P[ bini outcome1 ] outcome2 ] So, every observation falls into one of the bins of each predictor variable in the leaf node scorecards, and the overall score for each observation is simply the sum of the WOE values from those bins, normalized to a common scale. Scorecards that use WOE values based on these simple bin counts are referred to as naïve Bayes models. These models are used in ART because not only are the WOE values highly interpretive and robust when used as score weights, they can also be calculated very quickly using simple counts. This is particularly useful in ART, where many thousands of score weights are calculated in a typical run. Once the parent scorecard is created and the binned predictor variables are defined, the next step is to use ART to identify a set of binned candidate splitter variables. These can consist of all the variables considered during the construction of the parent scorecard, including those that made it into the model for the overall population. Given the binned predictors and the binned candidate splitters, an ART run can be launched. An ART run consists of a number of generations, each of which contains a fixed number of trees. In the initial population (i.e., generation #1), the trees are randomly created from the user-defined list of splitters and split points, along with a set of parameters that constrain the total number of leaves. In each leaf of each tree, ART constructs a naïve Bayes scorecard using the binned predictors that appeared in the parent scorecard. A score is generated for every observation by summing the WOE values associated with the bins in the scorecard of the leaf that observation fell into. 2006 Fair Isaac Corporation. All rights reserved. 3

Given the scores for each observation, the measure of predictive power used by ART to evaluate the effectiveness of each system of scorecards is Divergence, which is calculated as follows: Divergence = 1 ( P( S outcome ) P( S outcome )) log ds 1 2 e P( S outcome ) P( S outcome ) Divergence simply measures the ability of the scores from a given tree to separate the two outcome classes, where higher values indicate a greater degree of predictive power. In ART, this Divergence value calculated from the system of scorecards appearing in the leaf nodes of the segmentation trees serves as the fitness value that drives the genetic algorithm search. After Divergence has been calculated the genetic algorithm takes over. Below is a snapshot of the genetic algorithm search process in ART. 2 The genetic algorithm first ranks the trees by Divergence and identifies an elite set, which are the best N trees by this measure of fitness. This elite set is automatically promoted to the next generation of trees. This is important because it serves as the memory of the search algorithm, always retaining the best solutions found so far. Next, the genetic algorithm uses a process called fitness proportionate selection to choose pairs of parent trees that will be mated to create new child trees. Fitness proportionate selection leverages the Darwinian concept of survival of the fittest, whereby trees with higher Divergence are more likely to be selected for mating. The process itself is akin to a roulette wheel, where each tree is assigned a slot on the wheel, and the relative sizes of the slots is proportional to the relative divergence of the trees. Following fitness proportionate selection, the selected pairs of parent trees are then mated to create new child trees that will comprise the subsequent generation. Mating is accomplished by randomly selecting a node in the mother tree and replacing its entire sub-tree with the sub-tree selected from the father tree. This branch swapping procedure is then checked for logical consistency, and any empty or illegitimate nodes are pruned back up to the lowest valid parent node. 2006 Fair Isaac Corporation. All rights reserved. 4

The final step in the creation of the next generation of trees is called mutation. It involves randomly choosing a subset of the newly created child trees and replacing one or more of its splitters and split points with one selected from the user-defined list of possible candidate splitters. This process of mutation helps to maintain diversity in the search, and in many applications, this diversity is the key to ensuring a near-optimal final solution. This entire process is repeated over successive generations of trees until the values of Divergence stop improving on a holdout data sample. At that point, ART will return to the analyst the segmentation logic and leaf-node scorecard details of the very best overall set of trees. ART in practice Adaptive Random Trees is a breakthrough technology that has enabled the application of the segmented scorecard methodology across a wide range of applications. As the inventor of this technology, Fair Isaac has benefited from the massive scalability enhancements enabled by this unique and highly focused automated data mining tool. Our clients have also benefited from much more effective segmentation solutions enabled by a smarter tree search technology, delivered with greatly reduced turn-around times. As we continue to expand our genetic algorithm-based suite of predictive modeling power tools, we look forward to continuing our work with our clients and partners to deliver more world-class decision management solutions. ART availability Segmentation ART is now available as an optional module for the Model Builder scorecard platform. Designed by seasoned modelers, Segmentation ART overcomes the challenges that have made segmented model systems impractical for many analytic development projects. Segmentation ART technology leverages the capabilities of Model Builder scorecard for importing data, visualizing and exploring predictive patterns, defining predictive variables, creating models, evaluating their quality and swiftly deploying predictive analytics. By automating the process of identifying optimal segmentation, Segmentation ART delivers substantial benefits to your model development process. The ART process results in more elegant trees. These trees often reduce the number of models required, and dramatically cut development, deployment and maintenance costs. ART segmentation helps you boost the precision of your models, enhance the speed of the model development process, improve efficiency and reduce costs. 2006 Fair Isaac Corporation. All rights reserved. 5

About Fair Isaac Fair Isaac (NYSE:FIC) makes decisions smarter. As the world leader in decision management solutions driven by advanced analytics, Fair Isaac unlocks value for people, businesses and industries. Companies worldwide use Fair Isaac technology to make billions of faster, more profitable decisions a year in credit management, marketing, fraud, collections, bill payment and other areas. The world s leading banks and credit card issuers rely on Fair Isaac technology, as do insurers, retailers, telecommunications providers, healthcare organizations and government agencies. Founded in 1956, Fair Isaac serves thousands of clients through offices in nine countries, and helps millions of individuals improve their credit health through the www.myfico.com website. Visit Fair Isaac online at www.fairisaac.com. Fair Isaac Corporation (NYSE:FIC) provides the leading solutions and technologies for Enterprise Decision Management, giving businesses worldwide the power to make smarter decisions. Corporate Headquarters: Offices Worldwide: 901 Marquette Avenue, Suite 3200 Brazil, Canada, India, Japan, Minneapolis, MN 55402 Malaysia, Singapore, Spain, 1 800 999 2955 from the US United Kingdom, United States 1 612 758 5200 from anywhere info@fairisaac.com email www.fairisaac.com Fair Isaac and It s just a smarter way to do business are registered trademarks of Fair Isaac Corporation. Other product and company names herein may be trademarks or registered trademarks of their respective owners. 2006 Fair Isaac Corporation. All rights reserved. 2092WP 04/06 PDF