INFORMATION FUSION FOR
|
|
|
- Roger Gilmore
- 10 years ago
- Views:
Transcription
1 INFORMATION FUSION FOR IMPROVED PREDICTIVE QUALITY IN DOMAINS WITH HIGH INFORMATION UNCERTAINTY By RIKARD KÖNIG SCHOOL OF HUMANITIES AND INFORMATICS UNIVERSITY OF SKÖVDE BOX 408, SE SKÖVDE, SWEDEN SCHOOL OF BUSINESS AND INFORMATICS UNIVERSITY OF BORÅS SE BORÅS, SWEDEN DEPARTMENT OF TECHNOLOGY ÖREBRO UNIVERSITY SE ÖREBRO, SWEDEN A trend is a trend is a trend, but the question is, will it bend? Will it alter its course through some unforeseen force and come to a premature end? Cairncross (1969)
2
3
4
5 Abstract Data and information is today central to all organizations and is gathered from numerous sources in volumes that succeed any decisions makers ability to comprehend. Thus the need of decision support system is today greater than ever and is becoming a crucial part in many organizations. Decision making techniques create models of the collected data according to some algorithm; the model can then be used to explain patterns in the data or to predict the outcome of some future event. However the choice of decision choosing technique is not easy as there is no single technique that outperforms all other when evaluated over a large set of problems. A technique can also produce different models for the same data which makes the evaluation of the performance even more cumbersome. Instead an ensemble of several models sometimes created with different techniques is often used since ensembles have been proven to be superior and more robust than each single part by its own. Yet this superiority of ensembles techniques does not mean that they always achieve a sufficient performance to facilitate an automatic system. Most often the reason to the inadequate performance is that the data is imprecise, contain large uncertainties or that important variables simply are missing. In these cases the ensemble prediction often has to be adjusted manually to achieve an actionable performance. Here the strength of combining several models turns to a disadvantage as the reason behind a certain prediction becomes hard or impossible to interpret. Without a basic understanding of the underlying reason behind a prediction a manual adjustment cannot be done in a rational way. In these situations the ensemble model could be all together abandoned in favor for a transparent model or a more comprehensible model could be extracted from the ensemble. The later approach is called rule extraction and has been shown to be superior to creating a transparent model from scratch. This thesis continuous previous work on a rule extraction algorithm called G- REX (Genetic Rule Extraction) that has been shown to outperform both transparent- and other rule extraction techniques in terms of both accuracy and comprehensibility. The results are both theoretical and practical improvements of G-REX that together results in a better decision support for situations where ensemble techniques fail to give adequate predictions. G-REX is improved with a novel simplification technique that further enhances the comprehensibility and of the extracted rules and an improved technique for case based probability estimations. Probability estimation can be used to identify which cases that needs special attention from the decision maker and which that can safely be handled by the
6 ensemble. G-REX credibility as a rule extraction technique is also further strengthened by studies showing its suitability for estimation tasks an evaluation against the rule extraction criteria consistency. Finally it is also shown how tailoring of the representation languages and the performance metric used for optimization, can further improve G-REX performance.
7
8 Contents 1 Introduction Problem Statement Objectives Thesis outline Information fusion JDL-Model OODA-Loop The OODA-loop in relation to the retail domain Knowledge Discovery in Databases Data Mining Data Mining Tasks Machine learning vs. Data Mining Statistics vs Data Mining Predictive modeling Data Missing values Outliers Preprocessing Normalization Evaluation Holdout sample Random subsampling Cross validation Comparing Models Wilcoxon signed rank test Friedman test Nemenyi test Error Metrics for Classification Tasks Accuracy (ACC)... 35
9 4.5.2 Area under ROC-curve (AUC) Brier Score (BRI) Error Metrics for Estimation Tasks Mean absolute deviation (MAD) Coefficient of Determination (r 2 ) Root Mean Square Error (RMSE) Mean Absolute percentage error (MAPE) Geometric Mean of the Relative Absolute Error (GMRAE) Datasets used for evaluation Classification datasets Estimation datasets Data Mining Techniques Linear regression Decision trees Creation Pruning Artificial neural networks Basics of a Neural Network Evolutionary algorithms Genetic algorithms Genetic programming Genetic Programming Ensembles Calculating member weights Ensemble creation Member selection Estimating ensemble uncertainty Related work Rule extraction Rule Size... 63
10 6.1.2 Genetic Rule Extraction (G-REX) Market response modeling Market response modeling using Neural networks Motivation of empirical studies Case studies Increasing rule extraction comprehensibility Summary Background Method Experiments Results Discussion Conclusions Considerations Using GP to Increasing Rule Quality Summary Background Method Experiments Results Conclusions Considerations Instance ranking using ensemble spread Summary Background Method Experiments Results Conclusions Inconsistency Friend or Foe
11 8.4.1 Summary Background Method Results Conclusions Discussion and future work Considerations Genetic programming A Tool for flexible Rule Extraction Summary Background Method Results Considerations ICA Summary Background Method Results Discussion Conclusions Conclusions Concluding remarks Future Work References
12
13 1 Introduction Today decision makers often have access to numerous information sources containing vast amounts of data and information; i.e. processed data. Often the data is collected using several sensors to gain as much information as possible about activities related to a decision. [BPR02] argues that this approach is not necessarily the best one to support a decision maker. First, all information may not be required to adequately execute the task. Secondly, the resources required to process all this information might exceed the human information processing capabilities. This problem is also recognized by [HL01] and [SHBB03] as they point out that a decision maker will need an automated or semi automated decision support system (DDS) to make use of large quantities of data, especially if the data is of an heterogonous nature. Another motivation for a DSS is that real data often is associated with an amount of uncertainty or imprecision which can be hard to utilize in an optimal way for a human decision maker. Information fusion (IF) is a research area that focus on achieving synergetic effects by combining different information sources. There are numerous definitions of IF for examples see [HL01], [CGCP05], [Das01], [HW01] and [GB05] but what they have in common is the emphasis on maximizing the useful information content acquired from heterogeneous sources. The International society for information fusion has adopted [Das01] definition of information fusion: the theory, techniques and tools created and applied to exploit the synergy in the information acquired from multiple sources (sensor, databases, information gathered by humans, etc.) in such a way that the resulting decision or action is in some sense better (qualitatively or quantitatively, in terms of accuracy, robustness, etc.) than would be possible if any of these sources were used individually without such synergy exploitation. [CGCP05] also includes reducing or managing uncertainty as an important task in their definition of IF. IF techniques are sometimes implemented in a fully automatic process or in a Human-Aiding process for Analysis and/or decision support [HL01]. Techniques based on IF are therefore a natural choice when designing decision support systems. Traditionally most IF research have been done on military applications but [SHBB03] notes that the civilian sector is showing increased attention. This is not strange as decision makers have the same 13
14 basic requirements of decision support. To be able to effectively manage ones assets it s crucial that a decision maker gets an accurate, immediate and comprehensible presentation of the current situation, which is often only possible if the available information is fused. An important task of IF is predictive modeling, i.e., the task of predicting a unknown (often future) value of a specific variable based on a model learned from known examples. Predictive modeling, which is perhaps the most used subfield of data mining (DM), draws from statistics, machine learning, database techniques, pattern recognition, and optimization techniques [HW01]. The importance of predictive modeling and data mining techniques for information fusion is recognized in both [GB05] and in [MC05] where its argued that data mining on high-dimensional heterogeneous data has become a crucial component in information fusion applications because of the increasing complexity of the collected information. DM and IF is closely related areas with the difference that DM focuses on raw data and IF concerns information. Another difference is that DM more often uses a single source where IF concerns multiple sources. With this said there is nothing restricting DM techniques to single sources. It is also noted by [GB05] that the DM area of pattern recognition (a foundation of predictive modeling) has a long tradition of finding synergy by using different information sources. Data Databases Sensors Sales Campaign Information Trend Z will affect X Competitor C performed action A Data Mining Data Fusion Information Fusion X will have a behavior similar to Y. Figure 1 - Relation between Information Fusion and Data Mining 14
15 As seen in Figure 1 DM is a foundation for IF but not all DM tasks and techniques would qualify as IF. A sub domain of IF is data fusion (DF) which is similar to DM as DF techniques uses raw data. The difference between the two domains is that DM most often uses past data stored in databases and DF focuses on fusing data taken concurrently directly from available sensors. A predictive model is created from variables that are thought to provide information of the behavior of the dependent (predicted) variable. These variables are called independent or causal variables and are assumed to have known constant values. If it is believed that the values of the dependent variable is related to its previous values it is seen as a special case of predictive modeling called time series analysis or forecasting. The primary task of time series analysis is to capture the relationship between current and past values of the time series. For predictive modeling the primary goal is to achieve high accuracy, i.e. a low error between the predicted value and the real value, when the model is applied to novel data. For certain problems, a simple model, e.g. a linear regression model, is enough to achieve sufficient level of accuracy. Other problems require a more complicated model, e.g. a non-linear neural network or ensembles of models. The choice of a highly complex model, however, entails that it is hard or impossible to interpret, a dilemma often called the accuracy versus comprehensibility tradeoff. Rule extraction, one approach to this tradeoff, is the process of transforming the opaque model to a comprehensible model while retaining a high accuracy. The extracted model can be used to explain the behavior of the opaque model or it can also be used to make the actual predictions. Rule extraction is based on the assumption that an opaque model can remove noise from the original dataset and thus simplify the prediction problem. Another issue is that accuracy is not always the most suitable performance metric due to several reasons as pointed out by [Faw06]. Some techniques for example, associate a probability estimate to its predictions and for some cases as when ranking customers, it could be more important to know how good these estimates are. In any case there is no single performance metric that is suitable for all problems and thus it is important to be able to optimize arbitrary performance metric when creating a predictive model. Controversially most techniques only optimize accuracy due to underlying algorithmic issues. Thus the choice of technique most often leads to optimization of a predefined error metric when it would be more natural to choose the performance metric best suited to the problem at hand. A similar same problem applies to the model representation as 15
16 most techniques only can optimize models of a predefined structure with specific operators. Again the choice of technique dictate the representation when it would be more suitable to tailor both the structure and operators to the problem. current Sometimes it is impossible to create a good model due to uncertainties, low precision, noise or simply because relevant data are missing. If a model lacks sufficient accuracy for a fully automatic system it is crucial that it is comprehensible as an manual adjustment otherwise cannot be done in a rational way [Goo02]. Other motivations for manual adjustments and there by comprehensible models are that all relevant data cannot always be presented in a structured manner, all data is not always available at production time or that a drastic environmental change could have occurred. It can be crucial to be able to explain the underlying rules of a prediction for legal reasons or so a human forecaster can learn from the found relationships [Goo02]. Finally most decision makers would require at least a basic understanding of a predictive model to use it for decision support [Goo02], [DBS77], [Kim06]. This suggests that future research should focus on the problems of designing methods and support systems that work with judgmental forecasters, enabling them to use their judgments effectively, but also encouraging them to make full use of statistical methods, where this is appropriate [Goo02]. A DSS should also support a decision maker by giving a case based estimation of the uncertainty of its predictions. Probability estimates are central in decision theory as it allows a decision maker to incorporate costs/benefits for evaluating alternatives [MF04]. 1.1 Problem Statement The previous description leads to the following key observations: It is often impossible to produce a predictive model that has a sufficient accuracy for a fully automated DSS. Hence most organizations make manual adjustments to mechanically produced prediction. Manual adjustments of prediction can only be done in a rational way if the prediction model is comprehensible. The most powerful techniques as ensembles are opaque which limits their usability when manual adjustments are needed. Rule extraction is an area which could bridge the gap between opaque accurate models and comprehensible models. This has been done successfully before but could be done in a better way when tailored to a certain problem domain. 16
17 Many predictive techniques are rigid and cannot be adapted to give a better decision support for a specific problem. DSS with the following properties are needed today: Ability to fuse heterogeneous data from several sources. Ability to produce explanations of models and predictions. Ability to produces instance based uncertainty estimations for predictions. Optimization of arbitrary performance metric. Ability to create models with a representation that is tailored to the problem at hand Objectives The problem statement above leads to the following objective: 1. Evaluate usability issues for predictive modeling in situations which suffers from imprecision, uncertainty and/or lack of relevant data. 2. Survey the literature for rule extraction techniques suitable for extracting comprehensible models from ensembles. 3. Evaluate the need of improvements for the rule extraction technique G-REX. 4. Analyzing the effects of optimizing alternative performance metrics. 5. Evaluate techniques for providing case based uncertainty estimations from ensembles. 6. Evaluating techniques for estimating variable importance and the effect of possible future and action. This thesis will address all objectives except 6 which instead will be a main focus in future work. 1.2 Thesis outline In the following sections the theoretical background will be presented starting with a more detailed description of information fusion. Next the fundamental concepts of knowledge discovery are introduced in section 3 which is followed by a discussion of different error metrics and statistical evaluation techniques in section 4. Relevant predictive modeling techniques are described in section 5 and related work is addressed in section 6. In section Fel! Hittar inte referenskälla. the theory and related work are related to each of the thesis objectives and theoretical and empirical gaps are identified to put the contribution of this work into context. Next the indentified gaps are addressed in six cases studies which are presented in section 8. Each cases study starts with a short summary which is 17
18 followed by theoretical background, a description of the method, experimental details, results and conclusions. More general conclusions for all studies are presented in section 9. Finally future work and open research issues are described in section
19 2 Information fusion Information fusion (IF) is about combining, or fusing, information from different sources in order to facilitate understanding or provide knowledge that is not evident from the individual sources. Closely related terms (or research areas) are data mining and knowledge discovery, two terms commonly used from an artificial intelligence (AI) perspective for the extraction of regularities from large datasets. [ANO05] Similarly, decision support constitutes another research area whose subject matter largely overlaps with information fusion, although with a stronger focus on the perspective of the human decision maker. Uncertainty management i.e. the reduction of overall uncertainty through the combination information sources is a very central task in Information fusion. 2.1 JDL-Model Data fusion is the foundation of information fusion and is defined in [SBW99] as: the process of combining data to refine state estimates and predictions. The JDL model was developed to describe the data fusion process and to simplify communication between data fusion applications, in the military domain. The JDL model consists of 5 levels which are described in [SBW99] as: Level0 Sub-Object Data Assessment: Estimation and prediction of signal/object observable states on the basis of pixel/signal level data association and characterization; Level1 Object Assessment: Estimation and prediction of entity states on the basis of observation-to-track association, continuous state estimation (e.g. kinematics) and discrete state estimation (e.g. target type and ID); Level2 Situation Assessment: Estimation and prediction of relations among entities, to include force structure and cross force relations, communications and perceptual influences, physical context, etc. Level 3 Impact Assessment: Estimation and prediction of effects on situations of planned or estimated/predicted actions by the participants; to include interactions between action plans of multiple players. Level 4 Process Refinement (an element of Resource Management): Adaptive data acquisition and processing to support mission objectives. All levels in the JDL-model are in some way related to the retail domain but for sale forecasting level two (situation refinement) and level three (threat refinement) are especially important. Retailers have access to large amounts of data but has problem with how to utilize the data to maximize the knowledge of the 19
20 relationships between the objects (products and customers). At level two of the JDL model, Retailers main problem is high complexity due to a huge number of individual objects (customers, purchases, products etc.). An example is how to identify the behavior of different products and group them based on campaign- or seasonal sensitivity etc. A variation is to fuse individual customers buying behavior into a typical behavior for a certain group of customers, to better be able to tailor customer specific campaigns. Level three is better described as predicting the result of a certain action in a specific situation. A concrete example at this level is to predict the impact on sales of different marketing campaigns which are dependent of the current situation, the campaign and competitors actions. Information fusion not only deals with actual fusion processes, but also with how the results of these processes can be used to improve decision-making. Decisions can be made based on fusion of information from past and present behavior as well as on projections of future behavior. [ANO05] Figure 2 - Information Fusion in the retail domain Figure 2 above is an example of IF in the retail domain. Large databases contain data about previous observations and actions. The data could contain information about sales, events, weather and own marketing campaigns. This data has to be fused with the current situation which contains stock level, current weather, events and competitors actions. Forecast models of future sales can then be based on this knowledge and used to plan future marketing and campaigns and order strategies. 2.2 OODA-Loop 20
21 Boyd s OODA Loop (Orient, Observe, Decide, Act) has a military origin and is to today central to IF and the dominant model for command and control. The loop is an iterative process that contains four tasks below which are presented in [Gra05]. Orientation: An interactive process of many-sided implicit cross-referencing projections, empathies, correlations, and rejections that is shaped by and shapes the interplay of genetic heritage, cultural tradition, previous experiences, and unfolding circumstances. Observe: The process of acquiring information about the environment by interacting with it, sensing it, or receiving messages about it. Observation also receives internal guidance from the Orient process, as well as feedback from the Decide and Act processes. Decide: The process of making a choice among hypotheses about the environmental situation and possible responses to it. Decide is guided by internal feed-forward from Orient, and provides internal feedback to Observe. Act: The process of implementing the chosen response by interacting with the environment. Act receives internal guidance from the Orient process, as well as feed-forward from Decide. It provides internal feedback to Observe. A unique and crucial feature for the OODA-loop is the tempo. In order to win, we should operate at a faster tempo or rhythm than our adversaries or, better yet, get inside the adversary's Observation-Orientation-Decision-Action loop The OODA-loop in relation to the retail domain The OODA-loop is relevant for the sales forecasting as the typical work process of marketing campaigns naturally follows the steps in the loop. The current situation (product behavior, stocks etc) has to known before a campaign can be planned. When the current situation and campaign plan is available predictions of the sales impact are generated and the stock levels can be planned accordingly. Finally the prediction model is evaluated weekly to facilitate correction of stock levels and early warning of forecast breakdown. The ultimate aim for a decision support system is to fully automate the decision process; there is no better decision support than to have the correct decision made for you. A fully automated system has two fundamental requirements. First the system has to have a sufficient performance and secondly the users have to trust 21
22 the system enough to use it and not to tamper with the decision. Figure 3 shows how different retail tasks fits into the OODA-loop. Figure 3 - The role of DSS in the OODA-loop If a DSS can achieve the required accuracy and gain the trust of its users it could be possible to fully automate the, orient and decide steps (the white area of the loop). This is the ultimate goal (surely an utopia) of this research. Observation is already semi-automated and will always acquire some manually work as it impossible to control competitors actions. Domain experience and some initial work have shown that we are far from this goal today, mostly due to poor data quality and lack of information. Until the data quality is greatly improved a DSS cannot be fully automated and can only give well founded advice. It should be noted that the current DDS systems are surely comparable or better than the average forecaster hunch but DDS of today generally lack the necessary trust of its users. Therefore gaining trust and improving accuracy should be main goals of future research. Accuracy could be achieved by increasing data quality or improving methods for how to best use machine learning (ML) techniques as neural networks or ensembles. Trust can only be gained by understanding why a certain action should be taken. It is therefore crucial that each prediction is accompanied by an explanation. This will also work as a way to educate the users in important factors affecting sales. Users could in turn contribute with business rules and initial knowledge to reduce learning time and increase accuracy. User contributions would also work as a way to increase the general trust of the system. 22
23 3 Knowledge Discovery in Databases Knowledge Discovery in Databases (KDD) is an interactive, iterative procedure that attempts to extract implicit, previously unknown useful knowledge from data [RG03]. There are several slightly different models of the KDD process but one of the more frequently used is CRISP DM [CCK+] which is a cross industry standard for KDD and defines a cycle of six phases. Business understandi Data understandi Deployment Data Data preparation Modeling Evaluation Figure 4 - Phases of CRISP-DM Business understanding. This initial phase focuses on understanding the project objectives and requirements from a business perspective, then converting this knowledge into a data mining problem definition and a preliminary plan designed to achieve the objectives. Data understanding. The data understanding phase starts with an initial data collection and proceeds with activities in order to get familiar with the data, to identify data quality problems, to discover first insights into the data or to detect interesting subsets to form hypotheses for hidden information. Data preparation. The data preparation phase covers all activities to construct the final dataset. Data preparation tasks are likely to be performed multiple times and not in any prescribed order. Tasks include instance and attribute selection as well as transformation and cleaning of data for modeling tools. 23
24 Modeling. In this phase, various data mining techniques are selected and applied and their parameters are calibrated to optimal values. Some techniques have specific requirements on the form of data. Therefore, stepping back to the data preparation phase is often necessary. Evaluation. An analysis of the developed model to ensure that it achieves the business objectives. At the end of this phase, a decision on the use of the data mining result should be reached. Deployment. The deployment phase can be as simple as generating a report or as complex as implementing a repeatable data mining process across the enterprise. KDD and data mining has sometimes been used as synonyms but over last few year KDD has been used to refer to a process consisting of many steps, while data mining is only one of this step[dun03]. According to this definition the data mining process is performed within in the modeling step of CRISP-DM. 3.1 Data Mining Data mining is often defined as the automatic or semiautomatic process of finding meaningful patterns in large quantities of data. The process needs to be automatic due to the huge amounts of available data and the found patterns needs to be meanings full in a way that gives some kind of advantage. Data mining is the core of modeling phase of CRISP-DM but is also strongly connected to preparation- and evaluation phases. The patterns found in a data mining session is given as a model also called a concept description as the model is supposed to describe a real world concept based on a set of examples of this concept All data mining techniques are thus induction based as they try to form general concept definitions by observing specific example of the concept to be learned. A concept is real phenomena such as the weather, customers buying behavior, house prices etc. A customers buying behavior could maybe be described by the attributes income, family size, weekday etc. An example is a recorded instance of the concept where the attributes has a specific constant value Data Mining Tasks Data mining problems are usually broken down to number tasks that can used to group techniques and application areas. Even if the number of task and the name of the task vary slightly they all covers the same concept, [BL04] define the following data mining tasks: 24
25 Classification. The task of training some sort of model capable of assigning a set of predefined classes to a set of unlabeled instances. The classification task is characterized by well-defined definition of the classes, and a training set consisting of preclassified examples. Estimation. Similar to classification but here the model needs to be able to estimate a continuous value instead of just choosing a class. Prediction. This task is the same as classification and estimation, except that the instances are classified according to some predicted future behavior or estimated future value. Time series prediction also called forecasting is a special case of prediction based on the assumption that values are dependent of previous values in the series. Times series methods are either univariate (one variable is forecasted based on its past realizations) or multivariate (when several variables are used to predict the dependent variable). Affinity Grouping or Association Rules. The task of affinity grouping is to determine which things go together. The output is rules describing the most frequent combination of the objects in the data. Clustering. The task of segmenting a heterogeneous population into a number of more homogeneous subgroups or clusters. Profiling or Description. The task of explaining the relationships represented in a complex database. These tasks can further be divided into two major categories predictive and descriptive tasks [BL04]. Classification, estimation and prediction are all predictive in their nature while affinity grouping, clustering and profiling can be seen as descriptive. Predictive task. The object of predictive tasks is to predict the value of a particular attribute based on values of other attributes. The attribute to be predicted is commonly known as the target or the dependent variable, while the attributes used for prediction is known as explanatory or independent variables. Descriptive tasks. Here the objectives are to derive patterns (correlations, trends, clusters, trajectories and anomalies) that summarize the underlying relationships in the data. Descriptive data mining tasks are often exploratory in nature and frequently require postprocessing techniques to validate and explain results. 25
26 Prediction is the most important task seen from a business perspective and especially for sales forecasting as it is the only task dealing with future values. However prediction is nothing else then a special case of classification and estimation and most of the techniques defined for these tasks can also be applied to Prediction. Hence this section will only discuss predictive tasks. Univariate methods are either not going to be considered in this study as the problem domain containing strong explanatory dependencies Machine learning vs. Data Mining Most data mining techniques comes from the field of Machine Learning (ML) According [Mit97] the field of machine learning (ML) is concerned with the question of how to construct computer programs that automatically improve with experience. The algorithms and techniques used within ML exploits ideas from many fields including statistics, artificial intelligence, philosophy information theory, biology, cognitive science computational complexity and control theory. ML also has many other applications than data mining such as information filtering and autonomous agents etc. However as this research is focused on how to combine different data sources and how to finding meaningful patterns only ML techniques applicable to data mining task will be covered Statistics vs Data Mining There are not a sharp line between statistics and data mining, some concepts such as decision trees (see section 5.2) have even been evolved independently in both fields. However there are a slightly different underlying approaches where statistical methods foremost are design for hypothesis testing while data mining methods are more focused on searching for the best among all possible hypothesis. [WF05] 3.2 Predictive modeling Predictive modeling is the task of building a concept model that expresses the target variable as a function of the explanatory variables. The goal of predictive modeling is to minimize the difference between the predicted and real values. A model representation consists of a set of parameters (attributes, operators and constants) arranged in some kind of structure. Predictive modeling is the process of tuning or training the parameters of the model using a data mining algorithm to fit a set of instances of the concept as good as possible. The instances that are used to build the model are consequently called the training set. A model can 26
27 have a predefined static structure or it can be developed dynamically during training. Operators Attributes Constants AGE < 23 True False DEFLATED Start = DIP True False DEFLATED INFLATED Figure 5 - Possible model of the balloon dataset How well a model fits a specific training set is defined as the difference or the error between the predicted values ŷ and the real value y and is calculated for training set by the use of an error metric. When the training error is sufficiently small the training stops and the model can be used to make prediction for new unseen novel data. Normally some of the training instances are set aside prior to training creating a test set which is the used to evaluate how good the model will perform on new novel data. Error Metric Model Representation Data Mining Algorithm Trained Model Predictions Training Data Test Data Novel Data Figure 6 - Predictive Modeling 3.3 Data A trained model is never better than the data used for training and it is therefore important to ensure that the right data is available before modeling start. If the domain is unknown it is important to consult domain experts to ensure that no potentially important attributes are missing, as no technique can model relationships that are not present in the data. In general it is better to include 27
28 more data than less as the data mining techniques are supposed to be able to find the important relationships and disregard superfluous attributes [BL04]. As mentioned above a dataset contains a set of instances which each consists of specific values of the attributes thought to describe the modeled concept. Table 1 below shows a dataset containing ten instances of the concept of inflating a balloon (a slightly modified version of the balloon dataset in the UCI repository[bm98]). There are four input attributes color, size, start condition, type of person and one target attribute, outcome which is the attribute which is to be predicted. Color Size Start AGE Outcome YELLOW SMALL STRETCH 25 INFLATED BLUE SMALL STRETCH 42 INFLATED YELLOW SMALL STRETCH 5 DEFLATED BLUE SMALL DIP 34 DEFLATED YELLOW SMALL DIP 7 DEFLATED? LARGE STRETCH 52 INFLATED YELLOW LARGE STRETCH 23 INFLATED BLUE LARGE? 3 DEFLATED YELLOW LARGE DIP? DEFLATED BLUE LARGE DIP 7 DEFLATED Table 1 - The balloons dataset There are two main types of attributes; numerical which describes attributes that can take a value within a certain range and nominal (also called categorical) where only a set of predefined values are allowed. A special case of nominal attributes are ordinal attributes which where the order among the predefined values matters. [TSK06] divide numerical attributes into two interval where the differences between the values are meaningful and ratio where both the difference and ratios are meaningful. Temperature recorded in the Celsius scale is an interval attribute as ratios is not meaningful; i.e. 2 degrees Celsius is not twice as hot as 1 degree. Examples of ratios attributes are length, weight, age and monetary quantities. In Figure 5 Color, Size, Start and Outcome are nominal attributes while Age is numerical- or more precise, a ratio attribute Missing values As is common in many real world datasets the data set in Table 1 contains several instances that have missing values (marked with?) for some attributes. Data can be missing for several reasons; the value was not recorded, the attributes 28
29 was not present for the instance or the value could missing due to some other error. Whatever the cause it is important to realize that the missing values have impact on the final model depending on how it is handled. Most data mining algorithms are based on the assumption that no data is missing and requires that missing values are removed prior to modeling. There are several strategies that could be used to remove missing values. Instances with missing values could be removed, missing values can be replaced with a certain value not present it the data or they can be replaced with a value that is representative for the dataset. However all strategies has its own flaws and which to chose has to be decided from case to case. A common method for continuous attributes is to replace the missing value with the mean value of instances with none missing values. In the same way nominal missing values can e replaced with the type vale (the most common class) Outliers Outliers are values that differ greatly from the normal distribution of an attribute. A common statistical approach is to model an attribute by fitting a Gaussian probability function to the data [TSK06]. A Gaussian model is fitted to a particular attribute by calculating the mean µ and the standard deviation σ of the attribute for. The probability of x belonging to that distribution is then calculated by (1) [WF05]. 1 f ( x) e 2πσ ( x µ ) 2 2 2σ = (1) Figure 7 shows the probability that a value belongs to a Gaussian distribution with a mean of 0 and a standard deviation of 1. The scale of the x-axis is in standard deviations and the y-axis shows the probability. Normally an outlier is defined as a value more than three standard deviations from the mean which correspond to 0.27% of all values. 29
30 Outliers can be removed, marked or replaced by a representative value. However it should be clear that an outlier is some kind of error and not just a rare value Preprocessing As seen above there are several aspects related to the data that has to be handled before modeling can begin. All process concerning data modification before modeling is known as preprocessing. Removing missing values and outliers are often followed of some kind of transformation of the data according to the need of the data mining technique. One common technique is normalization of the data to remove difference in scale of the attributes and another is dividing the data into training- and test sets Normalization Many data mining technique are sensitive to the scale of the attributes. If not handle attributes in a with scale will implicit be given a higher importance during modeling. This is easily handled by normalizing (2) the attributes to a certain range, often 0 to 1. If the max and min value is not known standardizing (3) can be used to transforms all attributes so they get zero mean and unit variance. Below X is the attribute to be transformed, x ' the new value of x, µ is the mean and σ the standard deviation Figure 7 - Gaussian distribution with mean of 0 and STD of 1 x min( X ) x ' = max( X ) min( X ) (2) x µ x' = (3) σ 30
31 4 Evaluation In predictive modeling a model is trained to minimize the error on a training set. However surprisingly often this error is a bad estimation of future performance [WF05]. Each data instances is given by some probability function and as long as the training set does not contain all possible instances it will not be 100% representative and thus the model cannot be 100% representative. For this reason the training error will always be optimistic in regards to the actual error for the whole population. A better approach is to use hold out sample or test set to estimate the error on unseen data and compare classifiers. However, the test will neither be representative but it will be unbiased and can therefore be used to estimate the true error. 4.1 Holdout sample A hold out sample is created prior to training by randomly removing a subset of the training instance and placing them in the test set. Normally two-thirds are used for training and one-third for testing. A downside of this technique is that there is less data to train the model which can result in an inferior model. On the other side if too little data is used for the test set it will contain a large variance making the estimated error less reliable. Another problem is that the class distribution can differ between the data sets and thus affect both the model and the estimation negatively. 4.2 Random subsampling And extension of the hold out sample technique is called random subsampling which simply repeats the process several times. The overall error is calculated as the average error for all classifiers. A Problem with this technique is that there are no control of how many times an instance is used for test and training. 4.3 Cross validation Cross validation is a more systematic approach to random subsampling. Instead of choosing the holdout set randomly several times each instances is randomly assigned to one of k equal sized subsets or folds before training. Next k models are trained each using one of the folds as test set and the other k-1 folds as training so that each fold is used as a test set once. In this way cross validation utilizes all instances are used once for testing and k-1 times for training. However the folds can still be unrepresentative as they are chosen randomly. One time consuming approach to this problem is a special case of cross validation called leave one out, where each fold contains a single instance. Leave one out uses the maximum amount of data for training the models and each test is mutual 31
32 exclusive but the method is rather unpractical for most problems. A better approach is to ensure that all folds have a representative class distribution when they are created. Stratification is a simple technique that does this by selecting (in a random manner) as many instances of each class for every fold. For cases where the number of folds and instances of a certain class do not add up a copy of an already selected instance can be used or the fold can be left as it is. Most often 10-fold cross validation with stratification is used to evaluate classifiers. 4.4 Comparing Models Sadly there are no free lunches in data mining, i.e. no technique can outperform any other over the space of all possible learning tasks [Sch94]. Hence, when performing data mining it is normal to create several models using different technique and setting. However this means that the models needs to compared in a structured manner to ensure that the best model is actually selected. There has been much research on the topic of the best way of to compare classifiers and numerous tests are available. However [Dem06] have reviewed the current practice and examined several suitable test both theoretically and empirically. They recommend two none-parametric tests, the Wilcoxon signed ranks test for comparison of two classifiers and the Friedman test with the corresponding posthoc test for comparisons of more classifiers over multiple data sets Wilcoxon signed rank test If only two models is compared over several datasets [Dem06] recommends the Wilcoxon signed rank test [Wil45], which is based on the ranked difference of the models. First the absolute value of the difference in performance di on each dataset is ranked from the lowest to the highest. In cases of equal performance the ranks of these datasets are split evenly, if there is an odd number one is ignored. Next the rankings are summed for the cases that give a positive difference R + and the negative differences R according to (4). + 1 R = rank( d ) + rank( d ) i d 0 2 i > di = 0 1 R = rank( d ) + rank( d ) i d 0 2 i < di = 0 i i (4) If the smallest of the sums T = min( R +, R ) are less or equal to the corresponding critical value (CD) in Table 2, the hypothesis that the models are equal in performance can be rejected. 32
33 N CD Table 2 - Critical values of the Wilcoxon test at 0.05 significance level For comparison containing more than 25 dataset (5) can be used instead of Table 2. Here the test can be rejected at a 0.05 significance level of 0.05 if z is less than z = 1 T N N N N 24 ( 1) ( + 1)( 2N + 1) (5) Friedman test A Friedman test [Fri37] is a nonparametric test for comparing a number of models over several datasets using their rank. Ranking is done by giving the best a rank of 1, the second best 2, etc. In case of a tie the average rank of the tying models are assigned. A models that are ranked after a tie is assigned a rank according to how many models that been ranked so far plus 1. The nullhypothesis for the Friedman test is that all models are equal in their performance and hence also their ranks R. Friedman statistic is distributed according to j with k 1degrees of freedom when N>10 and k>5. In cases with less models and fewer datasets exact critical values can be found in [Shes00]. χ ( ) ( + ) N 2 k k 1 F = Rj k k 1 j 4 2 χ F (6) + However, the Friedman test has been shown to be unnecessary restrictive and hence modified according to (7) with 1 of freedom (denominator). F F k (numerator) and ( k 1)( N 1) = N k ( N 1) ( 1) χ 2 F 2 χf degrees Some sample critical values for the F-Distribution is presented below in Table 3. The significance level is 0.05 and the numerator is given by the column and the denominator by the row. (7) DF
34 Table 3 - F Distribution Critical values for P=0.05 If only two models are compared the ANOVA test has more power if the requirements of an ANOVA test are fulfilled. However The ANOVA requirements are not always met and it has also been shown that ANOVA and the Friedman test in practice most often gives the same results [Dem06] Nemenyi test If the null-hypothesis of the Friedman test is rejected [Dem06] recommends the use of the Nemenyi test [Nem63] to analyze which of the models that differs significantly from each other. When all k models are compared to each other over N datasets, a difference in average rank greater than the critical difference CD, see (8) is significant. ( + 1) k k CD = qα (8) 6N Table 4 gives the value of q α depending of the significance level α and the number of models (k). #Models(k) q q Table 4 - Critical values for the Nemenyi test If only a single model needs to be compared to a set of models a more powerful test as the Bonferroni-Dunn test [Dun61] should be used. The only difference from the Nemenyi test is the values of q α of (8) see Table 5. As seen in the table the difference between the two tests gets greater with the number of classifiers. #Models(k) q q Table 5 - Critical values for the Bonferroni-Dunn test 4.5 Error Metrics for Classification Tasks 34
35 Error metrics are different mathematical function for calculating the error or difference of the predicted value ŷ and the real value y. Error metrics are used to select the best of several models. However as all metrics have slightly different properties it is important to ensure that the chosen metric actually measure the right thing. Accuracy is by far the most common error metric used for classification. However [PTR98] argues that using accuracy for evaluating classifiers has two serious flaws; i.e. it is presumed that the true class distribution is known and that the misclassification cost is equal for all classes. This is regarded as a problem as these assumptions rarely are true for real world problems. Instead the other metrics such as Receiver Operating Characteristic (ROC) curve or the Brier score is suggested as a more appropriate metric for comparing classification performance. The following section will therefore describe these metrics in turn Accuracy (ACC) Accuracy is simply nothing else that the percentage of all instances that are correctly classified. n ji 1 if y 1 i y i= = i ACC = where ji = n 0 if yi yi (9) As mentioned above it should be noted that accuracy assumes that the class distribution is known for the target environment. If this distribution is not known accuracy is a poor error as it is not possible to claim that the evaluated model indeed maximizing accuracy for the problem from which the data set was drawn [PTR98] Area under ROC-curve (AUC) ROC curves measure the relationship between hit rate and false alarm rate and are based on estimations of the probability that an instance belongs to a certain class. ROC has an especially attractive property in that they are insensitive to changes in class distributions. ROC graphs are commonly used in medical decision making, and have in recent years been used increasingly in machine learning and data mining research; see e.g. [Faw06]. In a ROC-curve diagram the y-axis expresses the number of true positives and the false positives are shown on the x-axis. Both axes show the percentage of all true and false positives to give the unit square the sum of 1. To draw a roc curve the instances is first ranked according to the probability that they belong to the 35
36 positive class. Next the curve is drawn by moving one step up for each instance that are a true positive and taking one step right if it is a false positive. True 100 Positives % 80% 60% 40% 20% False 0% 20% 40% 60% 80% Positives 100% Figure 8 - Sample ROC curve For a detailed description of how ROC curves are constructed see [Faw06]. To compare classifiers using ROC curves, the curve has to be converted into a single scalar value, normally the area under the ROC curve (AUC). Since ROC curves are projected onto the unit square, AUC will always be bounded by 0 and 1. An AUC of 1 means that the predictive model is perfect, while random guessing corresponds (the dotted line in figure Figure 8) to an AUC of Brier Score (BRI) A prediction made with a certain model can often be related to probability estimation. Because of the fact that for a single realization a probability estimation is neither correct nor wrong, probability estimations are verified by analyzing the joint (statistical) distribution of estimations and observations [SK02]. While AUC only evaluates how good the produced probability estimates are for ranking the instance the brier score suggest by [Brie50] evaluates how good these estimations really are. (10) shows the definition of the brier score (BRI) where n is the number of instances, r is the number of possible outcomes (classes), f ij gives the probability estimate that the instance i belongs to class j. Eij takes the value 1 if the event has occurred and 0 if it hasn t. 1 BRI = f E r n 2 ( ij ij ) (10) n j = 1 i = 1 36
37 A lower score is better and the lowest possible score is 0 which occurs when the all classifications are done correctly with 100% certainty. Random guessing from a binary problem with equal class distribution will result in a score of Error Metrics for Estimation Tasks Some error measures are more useful than others. However after examination [ARM01] notes that depending on the situation several error metrics could be relevant and argues for the benefit of presenting all relevant metrics. When comparing different estimation models over several datasets, it is important to use an appropriate error measure. For example if models are going to be compared over several datasets the error metric has to be relative. This is true for both estimations tasks [LZ05] and for time series prediction[arm01]. A relative measure is a measure which is relative to some reference. There are many choices for reference but usually some simple straightforward model like the mean is used. A relative measure is also important since values of the target variable can differ greatly in magnitude between data sets, which of course will affect the magnitude of the error. The reference also accounts for the difficulty of the dataset which also is of great concern in a comparison. It is not always worse to achieve a higher error on a difficult task then a slightly lower error on a simple task Mean absolute deviation (MAD) Mean absolute deviation is a metric simple to interpret and calculate. A problem with MAD is that it is sensitive to the scale of the predicted variable and thus is unsuitable when comparing models over several data sets. n ( yi yi ) i= 1 MAD = n Coefficient of Determination (r 2 ) (11) The coefficient of Determination r 2 is the proportion of the variability in the data that is captured by the evaluated model. r 2 is bounded between -1 and 1 where 1 is a perfect correlation and 0 signifies no correlation, values below zero corresponds to negative correlations. It should be noted that r 2 takes no account for the bias between the predicted and the actual value, i.e. even if a model has a perfect r 2 its prediction can still differ greatly in scale from the actual value [ARM01]. 37
38 r n ( yi yi )( yi y i ) i= = n 2 n 2 ( yi yi ) ( yi y i ) i= 1 i= (12) Root Mean Square Error (RMSE) RMSE is very common metric which is a bit surprising as it is well known that it is a bad choice for several reasons; it is poorly protected from outliers, has a low reliability and low construct validity [AC92]. It is not always wrong to use RMSE but care has to be taken especially when forecasting or comparing methods across series RMSE = i= n i= 1 ( y ) 2 i yi n (13) Mean Absolute percentage error (MAPE) Mean absolute error (MAPE) presented in (14) is only relevant for ratio attributes and also has a bias in that it favor low forecast [AC92]. A single high forecast error like y i = 150 when y i = 100 would give a MAPE of 50% while a low forecast of the same magnitude where y i = 50 and y i = 100 would result in a MAPE of 33%. However [Mak93] shows how this bias can be removed by a simple adjustment according to (15) which would give both previous forecasts an error of 40%. Other advantages of MUAPE it avoids problems when dealing with actual values close to zero and that no trimming is required. However is surely not as intuitive to interpret as the original MAPE. MAPE = n i= 1 y i n yi y i (14) MUAPE = y i yi 1 y + y / 2 n i= ( i i ) n (15) Geometric Mean of the Relative Absolute Error (GMRAE) In [AC92] several error measures were evaluated according to reliability, construct validity, sensitivity to small changes, protection against outliers, and 38
39 their relationship to decision making. Geometric Mean of the Relative Absolute Error (GMRAE) was presented as one of the strongest measures. GMRAE is based on Relative Absolute Value (RAE) which is defined in (16) where ys, m the prediction (forecast) made by model m on series s. RAE is relative to the random walk (rw) with zero drift, i.e. the most recent known actual value. For is series s, rw predicts ys 1 RAE i = y y s, m y y s, rw s s (16) Random walk is often used as a reference point since it is straightforward and easily interpreted. In addition, random walk most often outperforms the mean prediction for time series. To handle extreme values, the RAE is trimmed using Winsorizing according to (17) 0.01 if RAEs < 0.01 C= RAEs if 0.01 RAEs if RAEs > 10 (17) GMRAE summarizes the WRAE using the geometric mean. The arithmetic mean is not suitable as any arithmetic average will be dominated by its large terms. In other words, good small errors should be counted but would essentially be ignored. It might be reasonable to expect that a large error should be balanced by a sufficiently small error which is the case for the geometric average but not for the arithmetic averages [LZ01]. GMRAE = s 1/ s s= 1WRAE (18) 4.7 Datasets used for evaluation The experiments in this study are performed on dataset from a wide range of domains. All datasets are publicly available from the UCI Repository [BM98] and are frequently used to evaluate machine learning algorithms. The motivation of using these benchmark data sets is that they are all gathered from real domains and therefore will contain problems that a DSS may face in the real world. Below each dataset used in the experiments will be described briefly Classification datasets 39
40 Bupa liver disorder (BLD). The problem is to predict whether or not a male patient has a liver disorder based on blood tests and alcohol consumption. The first 5 attributes are all blood tests which are thought to be sensitive to liver disorders that might arise from excessive alcohol consumption and the sixth is the number of alcoholic beverages drunk per day. Each line in the file constitutes the record of a single male individual. Cleveland heart disease (CHD). Prediction of whether a patient has a heart disease or not. The attributes includes age, gender and measurement of eight relevant hart characteristics. Contraceptive method choice (CMC). This dataset is a subset of the 1987 National Indonesia Contraceptive Prevalence Survey. The samples are married women who were either not pregnant or do not know if they were at the time of interview. The problem is to predict the current contraceptive method choice (no use, long-term methods, or short-term methods) of a woman based on her demographic and socio-economic characteristics. Congressional Voting Records Database (CVR). This data set includes votes for each of the U.S. House of Representatives Congressmen on the 16 key votes identified by the CQA. The task is to classification if a congressman is a democrat or republican depending on these votes. Credit card application (CRA). All attributes names and values have been changed to meaningless symbols to protect confidentiality of the data. Final settlements (FSL). Final settlements in collective agreements reached in the business and personal services sector for locals with at least 500 members (teachers, nurses, university staff, police, etc) in Canada in 87 and first quarter of 88. German Credit data (GCD). Data related to 1000 German credit card application which among other things contains information about credit history, purpose of credit, amount, employment, age, etc. Glass Identification Database (GLA). This study was motivated by criminological investigation and concerns classification of glass using the proportion of the ingredients. At the scene of the crime, the glass left can be used as evidence, if it is correctly identified! Hepatitis Domain (HED). Prediction of the survival of patients with hepatitis depending on treatment and disease related measurements. Horse Colic database (HCD). Retrospective study concerning if the lesion for cases of horse coli was surgical or not. Attributes contain the age of the 40
41 horses, data from different medical diagnosis, in which hospital they were treated and if the horse survived. Iris Plants Database (IPD). The data set contains 3 classes of 50 instances each where each class refers to a type of iris plant. One class is linearly separable from the other 2; the latter are not linearly separable from each other. Johns Hopkins Ionosphere (JHI).This radar data was collected by a system in Goose Bay, Labrador. This system consists of a phased array of 16 high-frequency antennas with a total transmitted power on the order of 6.4 kilowatts. The idea is to predict good radar returns; i.e. those showing evidence of some type of structure in the ionosphere. Lymphography Domain (LYD). Detection of cancer using x-ray of lymph nodes and lymphatic vessels made visible by the injection of a special dye. This lymphography domain was obtained from the University Medical Centre, Institute of Oncology, Ljubljana, Yugoslavia. Pima Indians Diabetes Database (PID). Prediction of whether or not a patient shows signs of diabetes according to the World Health Organization criteria. All patients are females at least 21 years old of Pima indian heritage. Protein localization sites (PLS). A biological dataset concerning classification of the localization site of a protein. Attributes consists of measurements of cytoplasm, persiplasm and six membrane properties. Sonar, Mines vs. Rocks (SMR). This dataset contains instances obtained by bouncing sonar signals off a metal cylinder at various angles and under various conditions. The transmitted sonar signal is a frequency-modulated chirp, rising in frequency. Each attribute represents the energy within a particular frequency band, integrated over a certain period of time. StatLog (Heart disease) (SHD). Similar the Cleveland heart disease dataset but in a slightly different from. Thyroid disease (THD). The task for this dataset is to diagnose whether a patient is sick of hyperthyroid or not. Vehicle silhouettes (VES). The problem is to classify a given silhouette as one of four types of vehicle, using a set of features extracted from the silhouette. The vehicle may be viewed from one of many different angles. Waveform Database Generator (Version 2) Data Set (WDG). This is an artificial three-class problem based on three waveforms. 41
42 Wisconsin breast cancer (WBC). Prediction of whether a detected case of breast cancer is malignant or benign. Wine (WIN). Recognition of wine based on chemical analysis of wines grown in the same region in Italy but derived from three different cultivars. The analysis determined the quantities of 13 constituents found in each of the three types of wines. Zoo (ZOO). A simple database containing 17 Boolean-valued animal attributes. The "type" to be predicted is: mammal, bird, reptile, fish, amphibian, insect or invertebrate. For a summary of the characteristics of the dataset see Table 6. Size is the number of instances in the dataset. Classes are the number of output classes in the dataset. Num. is the number of numeric attributes, Nom. is the number of nominal attributes and Naïve is the proportion of the majority class, i.e. the accuracy achieved by only predicting the majority class. Naïve is presented a reference point to each dataset and shows the worst possible performance that a reasonable classifier should achieve. 42
43 4.7.2 Estimation datasets Dataset Size Classes Num. Nom. Naive BLD % CHD % CMC % CVR % CRA % FSL % GCD % GLA % HED % HCD % IPD % JHI % PID % PLS % SMR % SHD % THD % VES % WDG % WBC % WIN % Table 6 - Characteristics of classification datasets Auto price (AUT). The data set consists of car specifications from which the price of the cars should be predicted. There are three types of attributes that describes an instance: car specifications, assigned insurance risk rating and its normalized losses in use as compared to other cars. Boston Housing Data (BHD). The inputs variables is statistics related to the livings standard in the different suburbs of Boston. The target variable is the median value of owner-occupied homes in a certain suburb. Computer Hardware (COH). The problem concerns predicting the relative CPU performance based on six CPU properties, cycle time, memory size, etc. Diabetes Numeric (DIN). The objective is to investigate the dependence of the level of serum C-peptide on the various other factors in order to understand the patterns of residual insulin secretion. The response measurement is the logarithm of C-peptide concentration (pmol/ml) at the 43
44 diagnosis, and the predictor measurements age and base deficit, a measure of acidity. Pharynx (PHA). A dataset consisting of patients with squamous carcinoma of 3 sites in the mouth and throat (in the oropharynx). The objective of the study was to compare the two treatment policies with respect to patient survival time. Sleep (SLE). Includes brain and body weight, life span, gestation time, predation and danger indices for 62 mammals. The target is the total time the mammal spends sleeping. Veteran's Admin. Lung Cancer Trial (VET). The data set contains data about veterans with lung cancer. The input variables consist of information about the patients, the type of cancer and the treatment. The target variable is the patient s survival time. Wisconsin (WIS). Each record in the data set represents follow-up data for a breast cancer study. The target is the recurrence time of the cancer and the input variables is measures for the cell nucleus. Table 7 shows the general properties of the data sets with the same abbreviations that were used for the classification datasets. The mean value (Mean), the standard deviation (Std), the maximum value (Max), the minimum value (Min) and the number of outliers (Outl.) are all calculated for the target variable. Size Num. Nom. Mean Std Max Min Outl Naive AUT BHD DIA MAC PHA SLE VET WIS Table 7 - Characteristics of estimation datasets 44
45 5 Data Mining Techniques 5.1 Linear regression Linear regression is a simple method foremost suitable for numeric prediction that is frequently used in statistical application. The idée is to find the amount of how much each of the attributes a1, a2,..., a k in a dataset contributes to the target value x. Each attribute is assigned a weight wi and one extra weight is used to constitute the base level of the predicted attribute. x = w0 + w1a 1 + w2a wk ak (19) The aim of linear regression is to find optimal weights for the training instances by minimizing the error between the real and the predicted values. As long as the data sets contains more instances than attributes this is easily done using the least square method [WF05]. Linear regression is quite intuitive and easily understood but the downside is that they are poor when non numerical attributes are present and that they can t handle more complex nonlinear problems. [Dun03] 5.2 Decision trees Decision trees are machine learning techniques most suited for classification but also applicable for regression problems. In the data mining community decision trees has become very popular as they are relatively fast to rain and produce transparent models. A decision tree can be seen as a series of questions, arranged in a tree structure, leading to a set of predefined classifications. Most often a decision tree consists of nodes containing either a prediction or a boolean condition (the question) with two corresponding child nodes. The conditions is chosen in a way that split the data set used for training into smaller more pure set of records, i.e. sets of instances which are dominated of a single target class. When it s impossible to find a condition in a node that will make the resulting sets purer it s named a leaf node and labeled to predict the majority class. When used for classification the root node is first evaluated followed by the left or the right node depending on outcome of the condition. This process I repeated for each resulting node until a leaf node with a prediction is reached. A path from the root to a leaf can be seen as a rule of simple if else conditions. 45
46 plas> Negative insu> Negative Conditions Positive Leaf Nodes Creation A DT is built to minimize the classification error on a training set. The creation of the tree is done recursively by splitting the data set one the independent variables. Each possible split is evaluated by calculating the purity gain it would result in if it was used to divide the data set D into the new subsets S = { D1, D2,.., D s }. The purity gain is the difference in purity between the original data set and the subsets as defined in (20) where P( D i ) is the proportion of D that are placed in D. The split resulting in the highest purity gain is selected and the procedure are then repeated recursively for each subset to this split. s gain( D, S) = purity( D) P( D )*purity( D ) (20) There are several different DT algorithms ID3 [Qui86], C4.5 [Qui93], CART [BFO+83] and CHAID [Kas80] which all use slightly different purity functions. ID3 use information gain as purity function which is based on the entropy metric that measures the amount uncertainty in a set of data. Information gain is calculated for all possible conditions for all independent variables and the entropy ( E ) is defined in (21), where is the probability of randomly choosing an instances of class in the data set S. Figure 9 - Simple Decision Tree i= 1 i i i c 1 E( S) = pi log i= 1 p (21) i Entropy ranges between 0 and 1 and reaches a maximum when the probabilities for each class are the same Pruning When a DT fully grown it is optimized for the training set which often leads to an over fitting of the model and there by a high generalization error. Pruning is a 46
47 method that analyses a decision tree to make it more general by to removing week branches. Removing parts of a tree will of course result in a decrease in training accuracy but the idea is that it will make the tree more general and perform better on new unseen data. There are two approaches to pruning; prepruning that tries to stop the growth of the tree before week branches occur and postpruning where a fully grown tree is first created and then pruned. [WF05] notes that postpruning has some favorable advantages as for example a series of condition can be powerful together even they all are week by themselves. In general post pruning algorithms generates candidate subtrees which is evaluated on the training data or a new validation set containing previously unseen instances. Exactly how these candidate subtrees are created differs between algorithms but they all apply subtree replacement and/or subtree raising in some fashion. Subtree replacement start in the leaves of the tree and replaces the selected subtrees with single leaves. Subtree raising moves a subtree to a position higher up in its branch deleting intermediate nodes. During a pruning process a large number of candidate subtrees can be created and the tree with the best performance on the validation data is selected as the final tree. 47
48 5.3 Artificial neural networks Artificial neural network (ANN) is a Machine Learning (ML) technique loosely based on the function of the human brain. ANNs are extremely powerful in the sense that they are universal functional approximators and can approximate any function to any desired accuracy [Hor91]. MLP has been used to solve a wide variety of problems in and is frequently used for both classification- and regression tasks due their inherent capability of arbitrary input- output mapping [ZPH98] Basics of a Neural Network An ANN consists of a set of units (neurons) which are arranged in layers and connected by weighted links which passes signals between the units. The units in the first layer, the input layer, are connected to the input variables and receive inputs signals accordingly. When a signal is sent through a link it is multiplied with the weight of the link and thus decides the impact on the receiving unit. All incoming signals are summarized and an activation function is applied to calculate the responding output. Units in the input layer are connected to units in a hidden layer which can be connected to another hidden layer or to the output layer, i.e. see Figure 10. Input Hidden Output Figure 10 - A simple Neural Network Hidden Layers and hidden Units The number of hidden layers and hidden nodes in an ANN is a crucial design choice as they allow the network to detect the features and capture the pattern in the data by performing complicated nonlinear mapping between input and output variables. Without hidden nodes, simple perceptrons with linear output nodes are 48
49 equivalent to linear statistical forecasting models. One hidden layer is sufficient for ANNs to approximate any nonlinear function with any desired accuracy. Two hidden layers can sometimes be preferable as it can achieve the same result with less hidden units and approximate uncontinuous functions. A network with fewer units will be faster to train and will in principle generalize better. The most common way in determining the number of hidden nodes is via initial experiments. Several rules of the thumb have been proposed but none of these works well for all problems. Output layer and Output Units Selecting the number of output units is another important factor for the performance of the network. For regression problems a single output unit is sufficient and classification task are most often design with one of C coding where each class is represented by an output unit. Classification is done according to the associated class for the unit with highest output signal. There are two types of forecasting: one-step-ahead forecasting (which uses one output node) and multi-step-ahead forecasting which can be performed in two ways. The first is called iterative forecasting in which the forecast values are iteratively used as inputs for the next forecasts. Iterative forecasting only uses one output node. The second is called the direct method and uses several output nodes to directly forecast each step in the future. Finally there is also the cascaded method that combines the iterative and direct approaches. Even if the empirical results are not unanimous the direct approach is recommended as the iterative method will propagate an early error into all subsequent forecasts [ZPH98]. Activation functions The activation function determines the relationship between the inputs and outputs of a node and a network. In general, the activation function introduces a degree of nonlinearity that is valuable for most ANN applications. A small number of well behaved (bounded, monotonically increasing, and differentiable) activation functions are used [ZPH98]. These include: (1) The sigmoid (logistic) function: 1 (2) The hyperbolic tangent (tanh) function: 49
50 (3) The sine or cosine function: (4) The linear function: Different activation functions can be used in a network but usually only one type is used in each layer. For classification problems which involve learning about average behavior the logistic function is recommended. Forecasting problems, on the other hand involves learning about deviations from the average and for this the hyperbolic function is more suitable. For the output nodes the linear activation function is reasonable when the forecast problem involves continuous target values. It is important to note that feed forward neural networks with linear output nodes have the limitation that they cannot model time series containing trend. Hence, for this type of neural networks, pre differencing may be needed to eliminate the trend effects. However it is not clear if different activation function has a significant effect on the performance of the networks. [ZPH98] Training of a Neural Network A trained ANN is basically a mapping between input and output variables of the data. Training of the ANN is an optimization process of the connection weights. The most commonly used training algorithm back propagation is a gradient steepest descent method. A step size often called learning rate and a momentum parameter can be set before training can begin. These parameters play a critical role in the performance of the final network but there is no way to calculate the optimal learning rate and momentum for a given problem. Several experiments with different setting are often performed before selecting the final values. BFGS and Levenberg-Marquardt [ZPH98] and [HBD96] are two modifications of the back propagation algorithm that are more robust and have achieved significant improvements in training time and accuracy especially for forecasting problems. Another advantage of these methods is that the parameters used by the algorithms can be set automatically. 50
51 5.4 Evolutionary algorithms Evolutionary algorithms (EA) are a collection of algorithms and techniques inspired by Darwin s principles of natural selection and survival of the fittest. EAs are essentially methods for optimization based on the natural process which occurs frequently in nature when a species adapts to changes in its environment. There are four main EA paradigms: Genetic Algorithms (GA), Evolutionary Strategies (ES), Genetic Programming (GP) and Evolutionary Programming (EP). However, this section will not cover ES as it is similar to GA but less straightforward to apply. EP will also not be described as it is akin to GP but aimed at evolving artificial life. The main idée in EAs is to solve a problem using a population of competing individuals. The problem defines the constrictions of an environment in which the individuals (candidate solutions) are going to compete. If no prior knowledge exists about the nature of the environment the initial population is created randomly, as this will ensure an approximate average performance of the population as a whole. Each individual is then evaluated and assigned a score (fitness) that correlates to how god they fit the environment. Finally individuals are selected to be a part of a mating pool where they will exposed to genetic operations to create offspring for a new stronger generation. There are several different strategies for how the individuals are selected to be a part of the mating pool, two of the more common are roulette wheel- and tournament selection. In roulette wheel selection each individual is selected with a probability to its fitness. This can be pictured by a roulette wheel where each individual gets a slot where the size of the slot corresponds to its relative fitness. The structure of wheel stays the same for each selection allowing individual to be selected several times (reselection). Figure 11 - Roulette wheel selection Tournament selection is performed by selection k individuals randomly and then selecting the one with the highest fitness. A higher k will put a higher weight on 51
52 the fitness value while a k of one will disregard the fitness and choose individuals randomly. When creating a new population there are three fundamental genetic operations reproduction, crossover and mutation. Reproduction makes o copy of an individual, crossover splits each of two parents into two parts and recombines them into two new individuals and mutation makes a random change to an individual. Most EAs implementations are done in a so called steady state fashion where the genetic operations are applied to individuals in the mating pool until the new populations has the same size as the current population. New generations are continuously evolved in this manner until a certain number of generations are reached or until a good enough solution is found. The main difference between GA and GP is the representation of the individuals and how the genetic operations are applied; this will be described in detail in the following sections. 52
53 5.4.1 Genetic algorithms In Genetic algorithms, first presented in [Hol75], an individual or chromosome is represented by a bit string. Normally the size of the string is fixed to facilitate simple crossover operations. The initial population is created by randomly assigning 1 or 0 to each bit of each individual. A population can contain tenths, hundreds or even thousands of individual depending on the problem and how computational expensive the evaluation of an individual is. To evaluate an individual the bit string is first translated into a candidate solution and then evaluated in a fitness function. In GA a crossover is performed on two parents who are selected randomly from the mating pool. Next both parents are split into two parts at the same randomly decided crossover point. Two new offspring are finally created by combining one part from each parent in an orderly fashion which preserves the placement of each part, i.e. see Figure 12. Crossover point Parent Offspring Crossover 1 Parent Offspring Genetic programming Figure 12 - GA crossover Mutation is simply done by changing a random bit and reproduction makes an exact copy of the selected individual Genetic Programming In genetic programming (GP), introduced by [Koz92], an individual or program is represented in a tree structure similar to a decision trees, using predefined functions and terminals. Koza used the functions defined in the LISP programming language but there are no restrictions of which functions and terminals that should be used. However it is important that the function- and terminals set is chosen with the problem in mind as GP only can find solutions containing these elements. Selecting the wrong functions and terminals may result in a search space that doesn t include the optimal solution while superfluous functions will result in an unnecessary large search space. For data mining problems where one important aspect is comprehensibility, simple representation languages as Boolean- or If Else rules are most often used. 53
54 For Boolean rules the functions set F consists of and, or and the relational operators greater, less, and equals while the terminals set T is defined by the attributes of the current dataset and random constants. Table 8 shows the function set, the terminal set and the BNF for simple Boolean rules as the ones presented in Figure
55 F = [AND, OR, <, >, = ] T = [con,cat,c] exp ::= [ROp BOp] [BOp] [ROp BOp] [BOp] ::= [exp][and OR] [exp] [ROp] ::= [con] [> <] [C] [cat] [=] [C] [func] ::= > < [C] ::= Double in the range of associated [con] Class of associated [cat] [con] ::= Continuous variable [cat] ::= Categorical variable Table 8 - BNF for Boolean rules There are three main strategies for how the initial population is created in GP. Full creates trees by randomly selecting functions until a certain depth is reached where terminal is selected for all functions. Grow randomly select functions and terminals and can there by create unbalanced trees, in difference to Grow which always creates balanced trees where all terminals are defined at the same depth. The last strategy Ramped half and half creates half of the population using Grow and half using Full while at the same time varying the maximum depth. Ramped half and half is the most common creation strategy as creates a diverse population where the trees have different size and shape. In GP the crossover operation is performed by randomly selecting a crossover point in each tree and recombining the parts into two new individuals as seen in Figure 13. The crossover point can be different in each tree but the resulting tree has to be two legal programs. If two incompatible points are chosen the resulting programs are discarded and the procedure is performed again. 55
56 AND AND < OR = < Tal. 177 = > Cp 2 Old 3.1 Cp 1 Old 1.9 AND AND < = OR < Tal. 177 Cp 2 = > Old 3.1 Cp 1 Old 1.9 Figure 13 - GP crossover Mutation is done by randomly choosing a mutation point and deleting everything below this point. Next the Grow creation strategy is applied to create new part of the program starting at the same mutation point. Reproduction is done similar to GA by simply copying the program into the next generation. 56
57 5.5 Ensembles An ensemble is a combination of the predictions made by several different classification or regression models. It is well known that a combination of several models can improve the accuracy of the prediction on unseen data, i.e. decrease the generalization error. At the same time, an ensemble is by definition a set of models, making it very hard to express the overall relationship found in original variables. The option to use an ensemble thus is a choice of a black-box model prioritizing accuracy. Models can be combined by simply using majority voting (for classification) and by averaging (for regression) or by weighting in some fashion. [KV95] A combination of the models is only meaningful if they disagree to some extent. If all models would produce the exact same output no new information could be gained by combining them. the disagreement or the ambiguity of the ensemble member models for the instances, can be calculated using (22) where is the weighted ensemble average. [KV95] m ( ( ) ( )) 2 A = V x V x n ωm i i (22) i= 1 m If is the weighted average of the generalization error generalization error of the ensemble E follow (23)., the E = E A (23) This means that the accuracy of the ensemble will increase if the ambiguities can be increased without increasing the generalization error Calculating member weights There is no analytic way to calculate the absolute weights for the models in the ensemble. It is important to realize that a model can have such poor generalization that its optimal weight should be zero. In the forecast community simple averaging of the models is used. A problem with simple averaging is that it assumes that all models are equally good [HSY94]. One approach to overcome this assumption is to calculate the optimal linear combination (OLC) of the ensemble members trough its weights. Although various loss functions could be used the squared-error loss is most often used in practice. The objective is then to minimize the MSE using least square regression why the technique often is called MSE-OLC. For an overview of MSE-OLC see [HSY94]. Since each member is trained on the same data they are prone to co linearity problems which reduce the 57
58 robustness of the ensemble. An alternative approach is to allow one perceptron or a whole ANN to learn an adequate strategy to combine the networks[lcv02] Ensemble creation If the ensemble members are diverse but still have a comparable accuracy, calculating the optimal weights becomes unnecessary, as simple averaging will work. This is the assumption for all ensemble creation techniques which tries to create good ensembles members from the beginning. Two of the most common techniques is bagging [Bre96] and boosting [Sch89]. Bagging When bagging the ensemble members are created from different training sets formed by making bootstrap replicas of the original set. The bootstrap replicas are constructed by randomly selecting instances (with reselection) from the original training set. All replicas have the same number of instances as the original dataset but differ from it as they contain duplicates of some instances. To make a prediction for a regression problem the ensemble members are simply averaged. Boosting Boosting also creates different dataset replicas for each ensemble member but the creation of the replicas is not independent. All replicas are copies of the original training set but have different weights for each instance. An ensemble is created sequentially by adding one new member at a time. The weights are calculated depending on the performance of the current ensemble by giving instances wrongly classified a high weight and there by a high importance for the new member while the correctly classified instance receives a low weight. Boosting does not always help the training if the training data contains a lot of noise [Dun03] Member selection Adding more ensemble members to increase accuracy will only work until a certain point. The reason is that when the size of an ensemble increases the probability that a new member will increase the ensemble ambiguity decrease. Eventually adding more members will only result in unnecessary computations. Adding new poor members could in some cases even harm the ensemble. An alternative is to select a subset of candidate ensemble members. This could be achieved by considering the difference in MSE of the current ensemble with K members and a new candidate ensemble with K+1 members. The members are tried in order of their MSE starting with the member with the lowest error. A 58
59 new member should only be added to the ensemble if the following inequality is satisfied [LCV02]: + > + (24) 2 ( 2K 1) MSE[ f ] 2 E [ m m ] E m k new i new i new is the MSE of the ensemble with K members, refers to the mathematical expectation operator and is the error that should be produced by the new member. If this criterion is not satisfied the candidate member is discarded and the next candidate member in the ordered sequence is tested. The ensemble members could also be selected to maximize the diversity of the ensemble following [KV95]. Several different diversity measures exist and there are many different techniques for constructing the ensembles. For a survey of the most important and common measures and techniques see [Joh07] Estimating ensemble uncertainty One key advantage for ensemble methods is that it can provide case dependent estimates of prediction uncertainty. Estimating the uncertainty associated with a prediction is crucial if the prediction is to be used in a decision making situation [RS03]. The variety (or diversity) of the individual solutions is called the forecast spread or ensemble spread and in a way represents the forecast uncertainty [Zhu05]. The spread is mathematically defined as the standard deviation of the ensemble members from the ensemble mean [Zhu04]. A perfect ensemble should have a spread equal to the ensemble mean error (or high correlation between the ensemble spread and the ensemble mean error) [Zhu05], but there are no exact translation of what a certain spread will correspond to in uncertainty. This depends on the size of the ensemble error and the ensemble spread. Since the diversity is related to the ensemble uncertainty, it could be used to rank the confidence for predictions made for instances even in the test set. If the ensemble members disagree on a certain instance, it should be interpreted as a higher uncertainty for the ensemble prediction on that instance. A different approach to used in the weather domain is presented in [TCM01]. The predictions of the ensemble members are sorted into a number of bins, according to a fixed set of continuous ranges. The principle is that it should be equally probable that the predicted value would fall into any bin. In [TCM01] the bins are calculated using statistical weather databases. It should be noted that the statistics is only used to pick the boundaries and not to train the ensemble. Each instance has its own set of bins, but the ranges are the same for all instances. The 59
60 distribution of the predictions over the bins is calculated for each instance and each bin gets a score according to the number of predictions it contains. Two different approaches are used to rank the instances according to the bin scores. The first approach scores the instances according to the ensemble mode i.e. the highest bin score for each instance. A high mode corresponds to a compact ensemble where many ensemble members predict similar values. A lower mode indicates a diverse ensemble with little agreement between the ensembles. A higher value should presumably correspond to a lower prediction error. The second approach is member support i.e. the number of members who fall into the same bin as the ensemble prediction. 60
61 6 Related work 6.1 Rule extraction The most powerful predictive techniques as ANN or ensembles of different models are hard or impossible to understand for humans, i.e. they are opaque. Transparent models as decision trees are easily understood but often have an inferior accuracy. This becomes a problem when the task requires an accurate transparent model. As described above transparency or comprehensibility could be motivated by legal reasons, safety issues, user acceptance or the necessity of making rational forecast adjustments which is also argued by [Goo02], [DBS77], [CS96] and [ADT95]. The choice of using an opaque model with high accuracy or an inferior transparent model is often called the accuracy vs. comprehensibility tradeoff. One approach to overcome this tradeoff is rule extraction; i.e. the process of transforming an accurate opaque model to a transparent model while retaining high accuracy. Orignal dataset X1 X2 X3 T Trained opaque model X1 X2 X3 T Rule Extraction Extracted rule making prediction R R IF(X1<5 IF(X2< Figure 14 - Rule extraction A totally different motivation for rule extraction could also be that an opaque model exists and an explanation for its behavior is needed. A decision maker can choose if he wants to keep the prediction from the opaque model and only use the extracted model as an explanation or if he wants to make new prediction with the extracted model. Which options that should be used depend on the problem at hand. In general the opaque model will still have slightly higher accuracy but using its prediction will result in a less transparent solution as the extracted model only will explain the prediction to a certain 61
62 degree. For this reason the extracted model is also most often used for making the predictions as the extracted model should have retained a high accuracy and that the rule extraction was motivated by a need of comprehensibility in the first case. Rule extraction is and quiet mature field where much research have been performed and several different methods had been proposed. To facilitate comparisons of different rule extraction techniques and to set guidelines for further research [CS99] suggested five criteria that should be fulfilled by good rule extraction algorithm: Comprehensibility: The extent to which extracted representations re humanly comprehensible. Fidelity: The extent to which extracted representations accurately model the networks from which they were extracted. Accuracy: The ability of extracted representations to make accurate predictions on previously unseen cases. Scalability: The ability of the model to scale to networks with large input spaces and large numbers of units and weighted connections. Generality: The extent to which the method requires special training regimes or restrictions on network architecture. Some authors add the criterion consistency; see e.g. [TS93]. A rule extraction algorithm is consistent if rules extracted for a specific problem are similar between runs; i.e. consistency measures how much extracted rule sets vary between different runs. The motivation is that if extracted rule sets differ greatly, it becomes hard to put a lot of faith in one specific rule set. There are two basic categories of rule extraction techniques decompositional and pedagogical, and a third eclectic that combines the two basic approaches [ADT95]. Decompositional techniques try to extract rules by analyzing the internal structure of the opaque model (usually the hidden and output units of an ANN). An example of a decompositional technique is RULEX, for details see [AG95]. Pedagogical techniques (PT) treat the model as a black box and thus are applicable for all kinds of models. Pedagogical techniques convert the rule extraction problem to a learning task from the original inputs to the models outputs, i.e. the original target variable is replaced by the predicted value. Two examples of pedagogical techniques are VIA[Thr93] and TREEPAN [CS96]. Most pedagogical techniques could of course be applied directly on the original dataset as a normal classifier. However the underlying idée for this type of rule extraction is that it is assumed that the opaque model contributes in the rule 62
63 creation process by cleaning the original data set from noise. A data set with less noise should be a better foundation for rule induction. Orignal dataset X1 X2 X3 T X1 X2 X3 Prediction of rule R R Trained opaque model T Extracted rule R IF(X1<5 IF(X2< Rule Extraction dataset X1 X2 X3 M PT Figure 15 - Pedagogical rule extraction When rule extracting from ensembles only pedagogical rule extraction meets the generality criteria; i.e. it would be impossible to create a decompositional rule extraction technique that would be general enough to work on an arbitrary ensemble. As ensembles in general are more accurate and robust only pedagogical rule extraction techniques be considered in this research Rule Size If the guidelines given by [CS99] is to be followed it is not enough that the extracted model is transparent, it also needs to be comprehensible. A transparent model is a model that can be read by a human and understood in its part. A comprehensible model on the other hand needs to be understandable as a whole which is quite another thing. Figure 16 shows a Boolean rule written in prefix notation for the Pima Indians Diabetes dataset described in section 0 that is transparent but not very comprehensible. It does not help that each condition can be understood when there are too many conditions. There is not a hard line deciding whether a rule is comprehensible or not but when comparing two rules with the same representation the simpler (smaller) one is most often considered more comprehensible. This is also in accordance to old theorem called Occam s razor (other things being equal smaller theories are preferable to larger ones) and the minimum description principle (MDL)[Ris78]. Two rules rule created with the same representation can be compared in a straightforward way by simply counting the number of elements each rule 63
64 contains. If both rules have the same accuracy the one with fewer elements should be chosen. ( Or( And( Or( Or( Less( Triceps )( 827,609 ) )( Less( Serum )( 342,516 ) ) )( Or( And( And( Greater( Triceps )( 186,974 ) )( Greater( Plasma )( 133,614 ) ) )( And( Less( Plasma )( 302,766 ) )( Less( TPregnant )( 267,67 ) ) ) )( Or( And( Less( Pressure )( 155,735 ) )( Less( BodyMass )( 547,947 ) ) )( Or( Or( Greater( Pressure )( 505,055 ) )( Less( Triceps )( 288,807 ) ) )( Greater( TPregnant )( 574,294 ) ) ) ) ) )( And( Or( Or( Or( Greater( Age )( 480,969 ) )( Greater( Plasma )( 121,843 ) ) )( Greater( Triceps )( 692,08 ) ) )( And( Or( Less( Serum )( 219,791 ) )( Less( Triceps )( 373,068 ) ) )( And( Greater( BodyMass )( 229,002 ) )( Greater( Plasma )( 289,4 ) ) ) ) )( Or( And( And( Less( Triceps )( 131,931 ) )( Greater( BodyMass )( 78,237 ) ) )( Or( Or( Less( Age )( 304,78 ) )( Or( Less( Serum )( 141,117 ) )( Less( Serum )( 1,927 ) ) ) )( Less( Triceps )( 613,874 ) ) ) )( Or( Or( Greater( Plasma )( 82,681 ) )( Less( PedigreeF )( 459,841 ) ) )( Or( Greater( PedigreeF )( 753,176 ) )( Less( TPregnant )( 541,046 ) ) ) ) ) ) )( Or( Or( Or( Or( Greater( Age )( 204,502 ) )( Greater( PedigreeF )( 342,593 ) ) )( Or( Greater( Age )( 539,65 ) )( Greater( Pressure )( 810,805 ) ) ) )( And( And( And( Greater( BodyMass )( 722,018 ) )( Greater( Age )( 771,832 ) ) )( And( Less( PedigreeF )( 676,094 ) )( Greater( BodyMass )( 548,964 ) ) ) )( And( And( Greater( Plasma )( 496,494 ) )( Greater( TPregnant )( 626,094 ) ) )( Or( Less( BodyMass )( 202,682 ) )( Greater( PedigreeF )( 541,959 ) ) ) ) ) )( Or( And( And( And( Less( Plasma )( 586,935 ) )( Greater( Age )( 467,127 ) ) )( Or( Greater( PedigreeF )( 129,058 ) )( Greater( PedigreeF )( 39,673 ) ) ) )( Or( And( Less( Serum )( 536,914 ) )( Less( TPregnant )( 94,637 ) ) )( Or( Less( PedigreeF )( 606,103 ) )( Less( BodyMass )( 626,094 ) ) ) ) )( And( And( And( Less( Serum )( 598,499 ) )( Greater( BodyMass )( 510,324 ) ) )( And( Greater( Plasma )( 486,2 ) )( Less( TPregnant )( 722,018 ) ) ) )( And( And( Less( BodyMass )( 626,158 ) )( Greater( Serum )( 804,593 ) ) )( Or( Less( Plasma )( 266,305 ) )( Less( PedigreeF )( 845,997 ) ) ) ) ) ) ) ) Figure 16 - Transparent but not comprehensible model Genetic Rule Extraction (G-REX) One of the latest contributions to the Rule Extraction domain is a technique called G-REX (Genetic Rule Extraction) first presented in [JKN03]. G-REX is in its foundation a GP framework design to induce prediction rules for data mining tasks. As a rule extraction technique G-REX follows the methodology of pedagogical rule extraction algorithms closely. Several distinctive features in G-REX make it very suitable as a rule extraction algorithm for data mining tasks. G-REX main strength is that it was especially designed to bridge the accuracy versus comprehensibility gap and to facilitate tailoring of the rule representations to the problem at hand. G-REX as been shown to fulfill the rule extraction criteria comprehensibility, fidelity, accuracy, 64
65 generality and to some small extent scalability but has yet not been evaluated against consistency. Fitness functions According to Occam s razor, MDL and common sense a good prediction rule should be accurate and as simple as possible. This idée is translated straight into G-REX fitness function for classification by giving a reward for each correct prediction a rule makes and a small punishment for each element it contains, i.e. see (25). A parameter p is used to tune how much shorter rules should be favored but no exact size threshold is given. During evolution this will result in that each element in a rule will be weighed against how much accuracy it adds to the rule. k = n ( ( ) ( )) * (25) fitness = E i == R i size p r k r k r k = 1 For estimation tasks G-REX uses a fitness function based on the total absolute error (ae) made by the rule. To produce short comprehensible rules G-REX also uses a punishment for larger rules. The length punishment is calculated in the same way as for classification tasks and ae is negated as a higher fitness value usually corresponds to a fitter individual. k = n ( ) Regfitness = E( i ) R ( i ) size * p (26) r k r k r k = 1 Representation One of G-REX main strength is that it is possible to tailor the representation languages used for the extracted rule according to the problem at hand. However the basic classification model used by G-REX is a typical decision tree with binary splits based on a single condition. Optionally, G-REX also allows complex conditions where several variables can be combined using conjunctions and disjunctions; for a sample G-REX decision tree see Figure 17 below. Figure 17 Classification tree evolved by G-REX Similar to most regression trees, G-REX regression model consist of a normal decision tree with constant values as leaves. This model can also use complex 65
66 conditions and optionally allows comparisons between attributes. For a sample model, see Figure 18 below. Figure 18 - Regression tree evolved by G-REX Empirical results For classification tasks G-REX has been evaluated thoroughly in several studies i.e. [JKN03], [UCK+03], [JSN04], [JNK04], [JKN04] and [JKN05]. The opaque models used as target for the rule extraction have most often been ensembles of neural networks but have also included ensembles made of a mix of different models. Overall these studies show that G-REX consistently achieves a higher or comparable accuracy compared to standard decision trees algorithms as CART and C4.5 and Tree Fit defined in MatLab and also outperformed other popular rule extraction algorithms as TREEPAN and VIA. However the main result was not that G-REX was slightly more accurate than these algorithms but that it outperformed them with significantly more comprehensible (smaller) rules. G- REX is both an excellent rule extraction technique and good technique for creating accurate comprehensible classification rules. Although the rules produced by G-REX are often rather compact they still contain unnecessary elements on occasions. These elements can be parts of a rule that is never executed (introns) or simply logical nonsense. [Koz92] argues that introns are an important part of genetic programming since they help to protect valuable parts from being destroyed by the crossover operation. While introns are necessary in the evolutionary process they add meaningless complexity to the final rule. G-REX has not been thoroughly evaluated for estimation task but initial experiments demonstrating the concept was performed in with promising results in [JKN04]. G-REX used a simple representation language with constant values as terminals. 66
67 6.2 Market response modeling Normally sales forecasting is done using a market response model that is supposed to capture the factors that affect the market. Today most retailers have access to numerous information sources capturing sales and several controllable casual variables such as such as price, advertising and sales promotions. In addition, a market response model may also include uncontrollable variables such as temperature, lifestyle trends and competitor actions. A market response model should model the sales response function, i.e. how the casual variables affect the unit sales or the market share. If a sufficient performance of the market response model is achieved it can be used for planning and to simulate how different action related to the controllable variables such as price and marketing efforts, would affect future sales. It is important to realize that nonlinearity often arise naturally in response modeling. Most common is probably the relative decrease in the effect of certain market effort in relation to the investment. A marketing effort will always increase sales but each additional unit of marketing effort will give less incremental sales than the previous. Other nonlinearities that a sales response function can show includes minimum and maximum sales potential, saturation, threshold, and/or marketing mix interactions effects. [Zha04] Normally market response model are evaluated against naïve models such as random walk which does not include any casual variables. Random walk predicts the latest known value and can hence capture major trends. Hence for marketing response models to be superior in forecasting, there must be strong current effects from the causal variables, i.e. the marketing mix. Other conditions such as poor data quality, may blur the advantages of market response models over naïve models [Zha04]. Sales response modeling has traditionally been done using regression and econometric models but as seen in Figure 19 ANNs have been gaining popularity in the last fifteen years as they handle nonlinearity well and are general function approximators. Thus the following section will give a survey of the use of ANNs in market response modeling and how they perform in relation to more traditional methods. 67
68 Figure 19 - Number of publication regarding forecasting with ANN Market response modeling using Neural networks [AA99] compared neural networks and an econometric model for forecasting the sales of a certain notational brand of strawberry yogurt. Sales were recorded using hand scanner data and both manufactory and store coupons were tagged to corresponding purchase records. Other store specific activities as ad-features, instore displays, special prices and other promotions were recorded on the weekday with highest sale. Data was collected for 25 product categories but the study only focused on strawberry yogurt as it had the most variation in retailer activity and a substantially higher average number of sales. The neural networks used retailer advertising, price coupon value, the presence of special displays and in-ad coupons, average market share and lagged sales as inputs and total sales volume as output. Models with three or four hidden layers was used in the study and trained until the error on the validation set (20% of the data) started to increase. Several different sizes of the hidden layers were evaluated and the one with the smallest validation error was finally evaluated against a multiple regression model based on the same variables. Both models were evaluated using the 2 R correlation and the validation set. The best neural network model had three hidden layers and achieved an 2 R of 0.7 and a mean square error of 326 compared to the regression model that had =0.5 and MSE= However the best neural network model was chosen using the same validation set as was used in the evaluation, thus introducing a favorable bias for the neural network. Another study performed by [GKP99] compares neural networks with multiplicative competitive interaction (MCI) models for modeling market 2 R 1 Created with ISI Web of knowledge using the search string: ( neural networks OR ANN) AND ( time series OR forecast*) in Topic. 68
69 response. MCI models are the most widely used market share model and has been shown to be superior to simpler linear or multiplicative models [KH90]. A MCI model gives an estimation of the market share for each brand included in the model. However estimating an MCI model often requires vast amounts of data due to the large number of parameters that needs to be estimated. Market response modeling also requires that the parameters are estimated simultaneously du the effects of multi period promotions and advertising which often overwhelms the number of observations in the available data. The experiments were performed on the Information Recourses Inc. (IRI) coffee dataset containing 58 weeks and the catsup dataset collected by A.C. Nielsen containing 156 weeks. Each datasets contained data aggregated from several stores for small set of important brands which together represented more than 90% of the market. The target was to predict the market share for each brand using the casual variables, price and lagged market share; the catsup data set also included information about feature and display type. One ANN model with three hidden layers were trained for each brand and dataset and used to predict the market share in a one-step ahead rolling fashion. When the different models were evaluated on a 10 week holdout sample for the coffee dataset, the ANN models had a lower MAPE for all seven brands. However for the brand with the smallest market share one of the model could achieve a really good result. For the catsup dataset the ANN models were only superior for the biggest brand. The author suggests that it some instance could be favorable to use an MCI models but that these examples are rare in practice and thus neural networks offer a powerful tool for market share forecasting. In [Hrus93] the market response function for a dominating Austrian consumer brand was modeled with both neural networks and econometrics models. As the brand totally dominated the market competitor actions could be ignored and the response in actual sales could be modeled instead of market share. The dataset contained 60 monthly observations of the actual sales and the casual variables price, advertising budget, lagged advertising budget and the average monthly temperature. Several linear regression models were constructed and optimized using the least square procedure of Chocrane-Orcutt for first order autoregressive residuals but only the best was compared to the ANN model. The neural network had on hidden layer with four units (one for each input variable. When evaluated the ANN model clearly outperformed the regression model in regards of mean square error and the absolute deviation. It should be noted that 69
70 this study only tried to identify the market response function and did not evaluate models forecasting ability. Since the neural network was quit small and only contained a single hidden layer it was possible to make some plausible economic interpretations of the network weights. Finally it should be noted that even if [Zha04] shows many successful application of neural networks they also recognize the downside of neural networks being incomprehensible. 70
71 7 Motivation of empirical studies This section will relate the theory and related work presented above to the objectives and highlights any gaps that need to be filled to accomplish all objectives. In this way the contribution of this study will be made clear and put in context with theory and related work. 1. Evaluate usability issues for predictive modeling in situations which suffers from imprecision, uncertainty and/or lack of relevant data. The success of predictive modeling is inherently dependent of the quality of the data. In situation where the data are imprecise, uncertain and/or lack relevant variables there is not always possible to achieve an acceptable performance for an automatic system. Yet a decision maker still needs the best possible support to make a well founded decision. On one hand ensembles are a good choice in these cases since they are more robust and facilitate probability estimations; on the other hand ensembles are opaque which makes it hard for a decision maker to a use an ensemble prediction as decision support in a rational way. However rule extraction has been named as an approach that can counter the drawback of ensembles opaqueness. The probability estimations can be used to identify the cases which needs to be adjusted by a decision maker and the extracted rules could then be used as a decision support. Nevertheless which rule extraction technique that is best suited for this task is still an open question. 2. Survey the literature for rule extraction techniques suitable for extracting comprehensible models from ensembles. Only pedagogical techniques should be considered when extracting rules from ensembles as they are the only kind that is general enough to deal with an ensemble of arbitrary members. VIA, TREEPAN and G-REX are three pedagogical techniques that could be considered. However G-REX has previously been shown to outperform both TREEPAN and VIA in regards of accuracy and especially in regards of comprehensibility and is thus a stronger candidate when choosing rule extraction technique. 3. Evaluate the need of improvements for the rule extraction technique G-REX. 71
72 G-REX has been shown to extract rules that are accurate and often very compact. However these rules sometimes contain unnecessary parts such as introns. If the rules could be simplified by removing the unnecessary parts it would further improve the comprehensibility of the extracted rules. Note that since G-REX is designed to extract rules with arbitrary representation the simplification technique also needs to be able to simplify arbitrary representation. Study 8.1 presents and evaluates a novel technique to handle this problem. G-REX has not been evaluated against the rule extraction criteria consistency. This is a serious deficiency as G-REX is inherently inconsistent due to that GP is used for optimization. To increase the trust for G-REX as a rule extraction technique, study 8.4 evaluates G- REX in regards of consistency. Previous studies of G-REX have focused on classification tasks, and only a few experiments have been performed on estimation tasks and then with a simple standard representation on applied to a single domain. Hence G-REX is evaluated using different representations for estimation tasks in study 8.5 and for time series prediction in study 8.6. The later study concerns sales forecasting in the retail domain which is characterized by many of the data problems that are addressed in this work. Here G-REX extracts rules from a mixed ensemble and uses data sets which both lack many relevant variables and contain imprecise data. 4. Analyzing the effects of optimizing alternative performance metrics. There are a multitude of performance metrics that all have different properties but there is none that is superior to all others. Hence it is strange that most predictive techniques are designed to only optimize a single metric. When choosing from several models one could of course use another metric than the one used for creation, but it should be natural that performance could be gained by optimizing the metric that is actually needed. Since G-REX is based on genetic programming the performance metric is coded in the fitness function which could easily be adapted to contain any metric. To shows the benefit of optimizing alternative performance metrics, study 8.2 compares rules optimized with three different performance metrics. 5. Evaluate techniques for providing case based uncertainty estimations from ensembles. Ensemble uncertainty estimations are crucial for decision makers as it allows a decision maker to incorporate costs/benefits when evaluating 72
73 alternatives. Normally the ensemble spread, i.e. standard deviation of the ensemble members from the ensemble mean, is used as an estimate of ensemble uncertainty. However much works on ensembles are done in the field of weather forecasting, such as the binning technique presented in section If the lack of historical databases could be handled this technique could prove useful for more general cases of predictive modeling. In 8.3 different adaptations of the binning technique are evaluated. 73
74
75 8 Case studies 8.1 Increasing rule extraction comprehensibility This section is based on the paper Increasing rule extraction comprehensibility [KJN06]. Note that some G-REX settings have been changed to be compatible with the latest version of G-REX Summary In this paper an extension to the rule extraction algorithm G-REX is suggested and evaluated. As noted in section rules produced by G-REX sometimes contains introns that add meaningless complexity to the extracted rule. In this study we suggest a novel extension to G-REX that simultaneously removes both introns and any other unnecessary code elements from the extracted rules. When trying to simplify an extracted rule G-REX starts by replacing the original target variable with the corresponding prediction from the extracted rule. This new dataset is then used by G-REX to evolve a new population of rules. Here, however, the fitness function is altered. A rule that fails to classify all training examples exactly as the initial rule is given a very large penalty, making it extremely unlikely that a faulty rule is going to survive to even the next generation. Another, much smaller, penalty is used to encourage shorter evolving rules. This fitness function all but guaranties that this second round of G-REX will produce a shorter rule. To improve the performance and speed of the simplification process the initial population is created using the extracted rule. More specifically; the population is divided into five equally sized parts, each subjected to different levels of mutation. One part starts with just exact copies of the original rule while the remaining parts start with from one to four mutation operations on each individual. Experiments on 16 datasets from the UCI repository shows that the novel simplification feature of G-REX succeeded in simplifying an extracted rule by removing unnecessary parts. As a matter of fact even rather short rules could often be further simplified. It should be noted that this process obviously can be used on a rule set with arbitrary syntax since it uses exactly the same procedure as the original rule extraction. This study strengthens the argumentation for G-REX as a part of a standard data mining technique. G- REX is able to extract very short and comprehensible rules while retaining a high accuracy. 75
76 8.1.2 Background G-REX has been evaluated in several studies [JKN03], [UCK+03], [JNK04], [JKN04], [JKN05] and [JLK+06] which all shows that G-REX can extract short accurate rules. The produced rules are significantly shorter compared to standard transparent techniques as CART and C5.0 while having a higher or comparable accuracy. However as mentioned in section rules produced by G-REX sometimes contains introns. Even if [Koz92] argues that introns are an important part of genetic programming since they help to protect valuable parts add meaningless complexity to the final rule. Obviously it would be possible to (automatically or manually) remove introns and logical nonsense from the extracted rule after evolution has finished. A more subtle issue is, however, the fact that rules often include perfectly logical and semantically correct parts that still do not affect the predictions from the extracted model. In this study we suggest a novel extension to G-REX that simultaneously removes both kinds of unnecessary code elements in the extracted rules. Rule with intron (Marked in grey) IF(X2=T) Same rule without intron IF(X2=T) IF(X2=T) IF(X1>5) 1 IF(X1>5) Method Two main experiments are performed to evaluate two different approaches for eliminating introns. In the first experiment an approach that tries to simplify the final extracted rule by removing introns. The second experiment investigates if a continuous fitness function will hamper the creation of introns. As G-REX is a rule extraction algorithm an opaque model is needed as the target of the rule extraction. This study uses an ensemble consisting of models created with five different data mining techniques; CART [BFO+83], two different multi-layer perceptron (MLP) neural networks [RM86], a radial basis function (RBF) neural network [BL88], and a probabilistic neural network (PNN) [Spe90]. Figure 20 Simple rule with introns Simplification G-REX standard fitness function is presented in section in equation(25) where size is the number of elements in the rule and p is a parameter that controls the amount of punishment a rule gets according to its size. The number of introns 76
77 that can appear in a rule is in a way related to the value of parameter p. A lower p value allows bigger rules and more introns as the rules only get a relative small punishment which easily can be countered by a increasing the fidelity with a single instance. Introns can be removed automatically by simply noting which parts of the rule that are actually used when the rule is evaluated on the training set and then removing the unused parts. Logical nonsense is a harder to remove as there could be numerous of different constellations that would have to be handled. It would be almost impossible to define all possible cases beforehand. Instead we propose that G-REX could simplify the extracted rule automatically by performing a second evolutionary process with the aim of removing both introns and logical nonsense. In this evolution the fitness function should reward rules that are shorter than the extracted rule while still making the same predictions. To facilitate the use of a fitness function with these properties a copy of the dataset used for the rule extraction is created with the exception that the predictions of the opaque model is replaced with the predictions of the extracted rule. This new dataset is then used in combination with a low value of p (0.01) to evolve a new simplified rule. Orignal dataset X1 X2 X3 T Simplification dataset X1 X2 X3 R G-REX Trained opaque model M X1 X2 X3 T Extracted rule R IF(X1<5 IF(X2< Simplified rule S IF(X1<5 0 1 Rule Extraction dataset X1 X2 X3 M G-REX Figure 21 - The simplification process 77
78 As the fitness function used during simplification (Equation (27)) has very small size punishment the fidelity towards the extracted rule is of the highest priority. R( i k ) is the prediction made by the extracted rule Rand Sr ( ik ) made by the simplified rule. k= n is the prediction ( ( ) ( )) *0.01 (27) simplificationfitness = R i == S i size r k r k r k= 1 When using the simplification fitness function the size punishment is in reality only effective when evaluating between rules with the same fitness. To improve the performance and speed of the simplification process the initial population is created using the extracted rule. More specifically; the population is divided into five equally sized parts, each subjected to different levels of mutation. One part starts with just exact copies of the original rule while the remaining parts start with from one to four mutation operations on each individual. I practice this also all but guarantees that the outcome of the simplification will be 100% faithful to the extracted rule. Continuous fitness A different approach is to try to reduce the number of introns present in the evolved population. If fewer introns are present during the evolution the extracted rule would of course contains less introns. The standard fitness function for G-REX was presented in (25) above the parameter p was used to control the size of the rules. However when G-REX rules are analyzed more closely they still often contain some introns. A solution to this problem is to use a fitness function with a more direct connection between the fidelity reward and the size penalty. Papagelis and Kalles suggested such a continuous penalty for a fitness function in [PK01]; see (28) below: k = n ( ) 2 ( ) ( ) * (28) 2 fitness = E i == R i r k r k k = 1 x size + x where x is an arbitrary large number. A smaller x means a bias towards smaller trees while a larger x prioritizes accuracy. In the study Papagelis and Kalles used x-values of 1000 and Here this continuous penalty function is evaluated again when applied in G-REX Experiments In the following section the outline for the experiments are described in detail. 16 publicly available datasets from the UCI repository [BM98] were used for evaluation. For a description of the datasets see section 4.7. Each dataset are first randomly divided into five partitions. Each partition is divided into 75% training instances and 25% test instances. G-REX is executed r 78
79 in a batch of ten consecutive runs and the rule with highest fidelity towards the ensemble is selected as the final rule. Three different fitness functions were tested, all rewarding high fidelity towards the ensemble but using different size punishments. Two of the fitness functions, Small and Large used G-REX normal fitness function, see equation (25). Small was defined to produce very short rules and hence had a p = 0.33, Large had a p = to allow longer rules. A third fitness function Cont. used the continuous penalty described in section with the parameter x set to Each fitness function was applied to each dataset and the produced rules were simplified using the procedure described in section above. Comparing sizes between rules extracted by G-REX and CART is not entirely trivial. For problems with more than two classes the representation is identical but for binary problems G-REX used a slightly different representation. The difference is that G-REX evolves Boolean rules while CART still creates a decision trees. Regarding the overall size this has no major implication tough. As an example; a test for CART like X1 < 5.2 would require only one node. The same test for G-REX is (< X1 5.0); i.e. three nodes. On the other hand the two classes are implicit in the G-REX representation while they constitute the leaves in the CART tree. From this it was deemed exact enough to measure the size for both G-REX and CART as the total number of nodes Results The tables below show the result for respective fitness function on each data set. Results for the ensemble are omitted as the focus lies on comparing comprehensible models. However the ensemble clearly outperforms CART seen over all data sets which is not surprising and according to theory. Note that the last row gives the mean result over all data sets. Dataset CART Large Small Cont. ACC R ACC R ACC R ACC R BLD CLE CMC CRA PLS GLA SHD JHI IPD PID
80 THD VES CVR WDG WBC WIN MEAN Table 9 - Accuracy on individual datasets (CD=1.09) All techniques achieve comparable accuracy (see Table 9) and a Friedman test (p=0.089) also show that there are no significant differences between the techniques. Table 3 shows the mean length of the original extracted rules and the resulting size when the same rules were simplified. When comparing the original rules it is obvious that Small and Cont clearly outperforms CART and Large. This is also confirmed by a Friedman test (p=2.5607e-008) and that the critical value for this experiment is Both Small and Cont produces significantly smaller rules than both CART and Large but there are no significance between Small and Cont. Dataset CART Large Small Cont. Size R Size R Size R Size R BLD CLE CMC CRA PLS GLA SHD JHI IPD PID THD VES CVR WDG WBC WIN MEAN Table 10 - Size of extracted rules (CD=1.09) The average size of the simplified rules in Table 11 is notably smaller for Small and Large when compared to the original rules. G-REX could remove large 80
81 amounts of introns and unused parts from all rules produced by the Large fitness function but even these simplified rules are longer than the original rules produced by Small and Large. Even the rules produced by the fitness function Small, which supposedly produces the shortest rules could be further simplified. None of the rules produced by Cont. could be simplified. A Friedman test yields p = e-006 which indicates that there are a statistical difference between the techniques also after simplification. When compared using average rank Small is significantly better than both Large and Cont. as the critical difference for this experiment is (0.792). Dataset Large Small Cont. Size R Size R Size R BLD CLE CMC CRA PLS GLA SHD JHI IPD PID THD VES CVR WDG WBC WIN MEAN Table 11 - Rule size after simplification (CD=.792) Figure 22 below shows a rule extracted for the CLE dataset using the Small fitness function. The size of the rule is 31 which are a bit longer than the average rule extracted with Small. It is not obvious if the rule contains introns as it seems logical in a manual inspection. In reality the grey elements are unnecessary for predicting the training instances. They could of course be used to predict new unseen instances but as these elements lack support from the training set there is no rational motivation for using them. 81
82 AND < OR Tal. 177 OR OR = AND AND OR Cp 1 > < < = > = Old 4.6 Tre. 178 Tre. 172 Tha 6 Old 1.9 Tha 3 Figure 22 - Extracted rule for CLE After a simplification of the rule in Figure 22, G-REX outputs the rule in Figure 23 which makes the exact same predictions for the training set. When the introns have been removed the rule has a size of 11while having the same performance as the extracted rule. AND < OR Tal. 177 = > Cp 1 Old 1.9 Figure 23 - Simplified rule for CLE Discussion All techniques achieve a comparable accuracy but differ significantly in the size of the produced rules. Even if the rules produced by Large could be greatly simplified they are still bigger the rules produced by Small and Cont. This is probably due to the fact that the fitness function allows bigger rules with large amounts of introns. Larger more complex rules should be able to achieve a higher accuracy but they also greatly increases the search space making it harder to find the best solution. As the number of generations was the same for all experiments it is not surprising that Large is slightly inferior in both accuracy and size. Before simplification Cont and Small performed comparable regarding both accuracy and size. Small were slightly better on both metrics but the difference was not statistically significant. Using the simplification process the smaller rules produced by Small could be made even smaller while Cont s rules couldn t be simplified at all. This is probably the result of Cont s fitness function where every 82
83 extra rule part results in a high penalty making it crucial that every new part adds extra fidelity to the rule. Cont s search of fit individuals seems to be less efficient then SHORT as it produces both longer and less accurate rules, possibly due to the fact that it hampers the creation of introns. Compared to CART even the extracted rules are significantly shorter while having a better or comparable accuracy. It should be noted that the size of the CART trees could be limited to a certain size by controlling the allowed depth of the created tree. On the other hand this parameter can t be set automatically and does just not in any way weigh accuracy against the size of the rule Conclusions G-REX succeeded in simplifying an extracted rule by removing unnecessary parts. As a matter of fact even rather short rules could often be further simplified. It should be noted that this process obviously can be used on a rule set with arbitrary syntax since it uses exactly the same procedure as the original rule extraction Allowing larger rules has in this case a negative effect especially considering that even the simplified Large-rules are longer than the original rules produces by the other techniques. If larger more complex rules are needed care must be taken that the resulting larger search space are explored during a sufficient number of generations. Cont and Small have comparable size and accuracy when used for rule extraction but the simplification process greatly reduces the size of the Smallrules. After simplification the rules produced with Small are significantly smaller than the rules produces with Cont which could not be simplified at all Considerations Care has to be taken so that the number of generations matches the size of the search space. If comprehensibility is of importance a simplification should always be done in sequence with the rule extraction. 83
84 8.2 Using GP to Increasing Rule Quality This section is based on the article [KJN08] Summary Rule extraction is a technique aimed at transforming highly accurate opaque models like neural networks into comprehensible models without losing accuracy. G-REX is a rule extraction technique based on Genetic Programming that previously has performed well in several studies. This study has two objectives, to evaluate two new fitness functions for G-REX and to show how G-REX can be used as a rule inducer. The fitness functions are designed to optimize two alternative quality measures, area under ROC curves and a new comprehensibility measure called brevity. Rules with good brevity classify typical instances with few and simple tests and use complex conditions only for atypical examples. Experiments using thirteen publicly available data sets show that the two novel fitness functions succeeded in increasing brevity and area under the ROC curve without sacrificing accuracy. When compared to a standard decision tree algorithm, G-REX achieved slightly higher accuracy, but also added additional quality to the rules by increasing their AUC or brevity significantly Background Several authors have discussed key demands on a reliable rule extraction method; see e.g. [ADT95] and [CS99]. The most common criteria are: accuracy (the ability of extracted representations to make accurate predictions on previously unseen data), comprehensibility (the extent to which extracted representations are humanly comprehensible) and fidelity (the extent to which extracted representations accurately model the opaque model from which they were extracted). The accuracy of an extracted rule set is evaluated similar to all predictive models; i.e. typically measuring error rates using either a hold-out set or some cross-validation scheme. Comprehensibility, on the other hand is harder to assess. In most studies, comprehensibility is measured as the size of the extracted representation. This is an obvious simplification, chosen with the motivation that it makes comparisons between techniques fairly straightforward. In addition, size could be calculated in a number of ways. Some typical choices (for tree structures) include number of questions, total number of tests, number of nodes and number of interior nodes. Furthermore, comprehensibility is in practice, clearly not linear, making all measures based on size slightly dubious. All in all, though, the choice to evaluate comprehensibility using the size of the model must be regarded as the most accepted. 84
85 In this study we use G-REX as a rule inducer instead of as a rule extraction algorithm. The only difference to the normal use of G-REX is that the target of the GP is the actual target value instead of predictions of an opaque model. There are two reasons for this approach, first the aim of this study is to evaluate two new fitness functions and for this an opaque model is an extra step that will only add unnecessary complexity to the study. The second, more important reason is that this approach also allows us to demonstrate the advantages of a rule inducer based on GP. Most rule inducers producing comprehensible models only optimize accuracy. This is a fundamental design choice that is built into the algorithms, making it hard or impossible to change the optimization criteria. GP on the other hand is very flexible and can optimize arbitrary functions. G-REX is an excellent GP technique to use for this evaluation as it allows fitness function to be added and combined as buildings block for more complex optimization criterions. Another G-REX strength is that it can use arbitrary representation languages facilitating a tailored solution consisting of an optimal representation language and an optimal fitness function for each problem Method The purpose of this study is to evaluate the effect of different fitness functions when using GP for rule induction. Although accuracy is the standard criterion when evaluating predictive models, there are several other criteria that could be used. Here, area under the ROC curve (AUC), and a special fitness function called brevity is evaluated. Brevity is aimed at producing compact rule sets, thereby increasing the comprehensibility. During experimentation, comparisons are made against both the standard G-REX fitness function, primarily optimizing accuracy and against Mat Labs decision tree algorithm Tree fit. One of G-REX main strengths is that an arbitrary fitness function can be used to optimize one or more criteria. In previous G-REX studies the fitness function here called ACCfitness has been used. In this study two new fitness functions AUCfitness and BREfitness are introduced. All three fitness functions are described in detail below. ACCfitness G-REXs original fitness function (equation(25) in 6.1.2) had a disadvantage in that the parameter p used to control the size of the extracted rules had to be set individual for each data set. This problem occurred because of that the size punishment was related to the number of correct classified instances. Two rules with the same accuracy would be affected more or less by the size punishment depending on the size of each dataset. When using G-REX as a rule inducer this would have to be considered a drawback as the decision maker would have yet 85
86 another parameter to tune. To counter this problem the fitness function was modified by dividing the reward from the number of correct classified instances with the total number of instance i.e. the accuracy of the rule, see(29) below. k = n ( E( i ) == R ( i )) k r k k = 1 ACCfitnessr = sizer * p AUCfitness [PTR98] argues that using accuracy for evaluating classifiers has two serious flaws; i.e. it is presumed that the true class distribution is known and that the misclassification cost is equal for all classes. This is regarded as a problem as these assumptions rarely are true for real world problems. Instead the Receiver Operating Characteristic (ROC) curve is suggested as a more appropriate metric for comparing classification performance. Specifically, ROC curves are insensitive to changes in class distributions. ROC graphs are commonly used in medical decision making, and have in recent years been used increasingly in machine learning and data mining research; see e.g. [Faw06]. ROC curves measure the relationship between hit rate and false alarm rate and are based on estimations of the probability that an instance belongs to a certain class. For a detailed description of how ROC curves are constructed see [Faw06]. To compare classifiers using ROC curves, the curve has to be converted into a single scalar value, normally the area under the ROC curve (AUC). Since ROC curves are projected onto the unit square, AUC will always be bounded by 0 and 1. An AUC of 1 means that the predictive model is perfect, while random guessing corresponds to an AUC of 0.5. Since G-REX uses GP to evolve rules, it does not, just like standard decision trees, provide explicit probability distributions for the classifications. The number of correct and incorrect training classifications is, however, recorded for each leaf in the tree, making it possible to produce probability distributions if needed. n (29) If(plas>177.82) Negative [42/59] If(insu>77.82) Negative [2/3] Positive [8/15] Figure 24 - Correct classifications / instance reaching node The relative frequencies are normally not used directly for probability estimations since they do not consider the number of training instances supporting a 86
87 classification. The Laplace estimate is a common technique to produce probability distributions based on the support [MD02]. Equation (30) below shows how probability estimates p is calculated when using Laplace. N is the total number of instances, C is the number of classes and k is the number of training instances supporting the predicted class A. p classa k + 1 = (30) N + C In Figure 24 above, the probability estimate for the lower right node would be calculated as 8/15 without Laplace and 9/17 ((8+1)/(15+2)) using Laplace. When employing AUCFitness, G-REX uses Laplace estimates for ranking the instances during the ROC analysis. Overall, AUCfitness balances the AUC and the length of the rule in the same way that ACCFitness balances accuracy and length. AUCfitness = AUC size * p (31) i i r BREfitness The fitness function BREfitness is aimed at optimizing rule brevity. Instead of just using the length of the rules as a base for punishment BREfitness evaluates how the rule is actually used when classifying instances. The idea is that a tree must not only be relatively small but also should classify as many instances as possible using only a few tests i.e. the rule brevity should be high. The overall principle is that a typical instances should be classified using a few simple tests, while more complex rules should be used for more atypical patterns only. Although brevity could be measured using different metrics, we suggest that brevity should be calculated as the average number of tests needed to be checked in order to classify all instances in the data set, i.e. the brevity for a rule set r when classifying the instances k=1..n is: BRE r k = n # conditionsk = (32) # instances k = 1 # conditionsk is the number of conditions that have to be checked to classify instance k. If the counts in Figure 24 are calculated on test instances, the corresponding brevity is: (59*1+3*2+15*2)/( )=1.2. Note that using this definition, brevity does not consider accuracy at all. A rule with the best (lowest) possible brevity (1.0) could classify all instances incorrectly. Brevity should therefore only be used to evaluate rules with comparable accuracy. Because of this, BREfitness functions uses accuracy as a reward and brevity as a punishment 87
88 (a lower BRE means that the rule has better brevity). The size of the punishment is adjusted, again using a parameter p. BREfitness k = n ( E( i ) == R ( i )) k r k k = 1 r = n BRE * p Tree fit The Tree fit algorithm in Mat Lab is a decision tree algorithm for regression and r (33) classification which is based on CART [BFO+83]. Tree fit produces a binary tree with splits based on the input variables. For continuous variables the relational operators < and > are used. Categorical variables are split using the = operator (variable = category list) which is true if the current variable is one of the listed categories. Gini diversity index (GDI) is used as splitting criterion and each node is set to contain at least five instances. GDI measures the class impurity in the node given the estimated class probabilities ( ) GDI is given by (34) below. p 2 * j t, j = 1,..., J (for J classes). ( ) 2 GDI 1 p * j t = (34) When pruned, the tree is first evaluated on the training data using 10-fold cross validation and then pruned to the minimal subtree that is within one standard error of the minimal cross validation error Experiments This study uses 13 binary datasets from the UCI repository which are described in detail in section 4.7. Missing values for continuous variables are replaced with the mean value of all none missing values for the variable. Categorical missing values are, similarly, replaced with the type value of the variable. When missing values have been handled, each data set is divided into ten folds which are stratified to ensure that each fold will have a representative class distribution. For evaluation, standard 10-fold cross validation is used in the experiments. G-REX In this study G-REX executes in a batch of ten runs for each fold. The rule with the highest fitness is selected as the winner and is then evaluated on the test set. All fitness functions use p=0.05, which should produce short and comprehensible rules. G-REX used the GP settings presented in Table 12 below. Number of generations 100 Population size 1000 Crossover probability 0.8 Mutation probability Creation type Ramped half and half j 88
89 Maximum creation depth 6 Length punishment (p) 0.05 Batch size 10 Table 12 - G-REX settings used in all experiments G-REX can use arbitrary representations languages but in this study the simple ifelse representation languages presented using Backus-Naur form in Table 13 was deemed to be sufficient. Rule ::= [If] [If] ::= If ([cond]) then ([exp]) else ([exp]) [cond] ::= [con] [func] [C] [con] [func] [con] [cat] = [C] [exp ] ::= [If ] [R] [func] ::= > < [C] ::= Double in the range of associated [con] Class of associated [cat] [R] ::= Class of target variable [con] ::= Continuous variable [cat] ::= Categorical variable Table 13 - Representation language The tree displayed in Figure 24 is an example of this representation; if an ifstatement is true, the left subtree is evaluated, if not the right. An instance is classified as the value of the leaf it reaches, following the if-statements. Evaluation For actual evaluation, all G-REX fitness function and Tree fit are evaluated using accuracy, AUC and brevity. These evaluations are done regardless of optimization criterion to investigate whether AUCFitness or BREFitness perform better than the standard accuracy fitness function. The results are averaged over all folds and reported individually for each dataset. Overall performance of the different techniques is compared using a Friedman-test. The tests are performed at the 0.05 significance level, so a p-values larger than 0.05 indicates that the null hypothesis (median is zero) cannot be rejected at the 5% level; a p-value smaller than 0.05 indicates that the results are significantly different. The Nemenyi test is used to decide if the performance of two techniques differs significantly different. If the mean rank of two techniques is higher than the critical difference CD the difference is significant Results In the following section the results of the experiments are presented. Each table shows a different metric for the same rule set; i.e. accuracy, AUC, and brevity are 89
90 calculated for one rule set per technique and dataset, but the results are presented in three different tables. All results are averaged values for ten stratified folds. Next to the evaluated metric the rank among the evaluated techniques is presented. Table 14 below shows the achieved accuracy for the different techniques on each dataset. Tree Fit ACCfit BREfit AUCfit Dataset ACC R ACC R ACC R ACC R WBC HCD CRA GCD PID CHD SHD HED JHI FSL BLD SMR CVR MEAN Table 14 - Average ACC over 10-folds (CD=1.21) As seen in the table all techniques produce comparable results. A Friedman test confirms this as it yields p=0.210 ACCFitness achieves slightly higher accuracy than the other techniques. Table 15 shows the AUC values for the same rule sets that were evaluated against accuracy in Table 14. AUCfitness clearly outperforms the other techniques using this metric, achieving the highest AUC on 9 of 13 data sets. The second best technique is ACCFitness followed by Tree fit and BREFitness. A p-value of 8.06e-5 and a CD-value of 1.21 which shows that both AUCFitness and ACCFitness is significantly better than Tree fit and BREfitness. 90
91 Treefit ACCfit BREfit AUCfit Dataset AUC R AUC R AUC R AUC R WBC HCD CRA GCD PID CHD SHD HED JHI FSL BLD SMR CVR MEAN Table 15 - Average AUC over 10-folds (CD=1.21) When evaluating the rule sets against brevity (Table 16), the results for Tree fit is not reported as the representation languages is slightly different. Dataset ACCfit BREfit AUCfit BRE R BRE R BRE R WBC HCD CRA GCD PID CHD SHD HED JHI FSL BLD SMR CVR MEAN Table 16 - Average BRE over 10-folds (CD=0.879) In Table 16 it is clear that BREfitness produces rules with much lower brevity. It is worth noting that the same rule sets also performed rather well when evaluated 91
92 using accuracy. It is natural that BREfitness does not perform as well when evaluated against AUC since it produces very short rule sets leading to quite rough probability estimations. A Friedman test yields p= e-005 and a Nemenyi gives a CD=0.879 which shows that BREfitness produces significantly smaller rule sets than the other techniques. Another interesting result is that AUCfitness is significantly worse than all other techniques when compared on brevity. AUCFitness shortcomings is probably explained using the same argument as above; i.e.to achieve a high AUC it is necessary to have good probability estimations, which can only be produced by more complex rule sets Conclusions This study had two objectives, to evaluate two new fitness functions for the G- REX rule extraction algorithm and to show the advantages of using GP as a rule inducer. Both fitness functions perform well as they retain the same accuracy as the original fitness function while optimizing yet another other criteria. AUCFitness produces rules that have high accuracy and a higher AUC than Tree fit and the other evaluated fitness functions. BREFitness uses the proposed comprehensibility measure brevity to create rules which classifies the typical instances with simple conditions and only uses complex condition for atypical instances. The fitness function outperforms the AUCFitness and ACCFitness regarding brevity without losing significantly in accuracy. The experiments show that when G-REX is used as a rule inducer it is clearly a better alternative than Tree fit and similar decision trees algorithms. G-REX achieves the same or slightly higher accuracy than Tree fit, but adds additional quality to the same rules by increasing their AUC or brevity. A decision maker could certainly benefit from this extra rule quality, especially by directing G-REX to produce rules optimized on the quality measure best suited for the problem at hand. Another strength of GP based rule inducers is that the representation language can be tailored for the problem and thereby possibly increase comprehensibility, accuracy or both. Overall the experiments show that G-REX is well suited for classification tasks and should certainly be considered when a high quality transparent model is needed Considerations G-REX should definitely be considered when choosing rule inducer as it can use a tailored representation language and optimize more than one performance metric. 92
93 8.3 Instance ranking using ensemble spread This section is based on the paper Instance Ranking using Ensemble Spread [KJN07] Summary This paper investigates a technique for estimating ensemble uncertainty originally proposed in the weather forecasting domain. The motivation for using ensembles for forecasting is that they are both more accurate and more reliable compared to single model forecasts. Another key advantage for ensemble forecasts is that it can provide case dependent estimates of forecast uncertainty. The diversity of the predictions made by the ensemble members is called the forecast spread or ensemble spread and in a way represents the forecast uncertainty. Spread is mathematically defined as the standard deviation of the ensemble members from the ensemble mean. The overall purpose is to find out if the technique can be modified to operate on a wider range of regression problems. The main difference, when moving outside the weather forecasting domain, is the lack of extensive statistical knowledge readily available for weather forecasting. In this study, three different modifications are suggested to the original technique. In the experiments, the modifications are compared to each other and to two straightforward techniques, using ten publicly available regression problems. Three of the techniques show promising result, especially one modification based on genetic algorithms. The suggested modification can accurately determine whether the confidence in ensemble predictions should be high or low Background The domain of weather forecasting is closely related to information fusion. Data is collected from multiple sources and forecasting systems have to deal with a high degree of uncertainty. The forecasts also have to be constantly updated as conditions change rapidly. A lot of research has been performed in the weather forecasting domain during the last decade, often with great success. The research has been focused on ensemble forecasting, which today is an integral part of numerical weather prediction [TCM01]. Using ensembles for forecasting is both more accurate and more reliable compared to single model forecasts. Another key advantage for ensemble forecasting is that it can provide case dependent estimates of forecast uncertainty [Ehr97]. This is considered a requirement for weather forecasts, but it is clearly an advantage in all decision making, since a forecast with associated uncertainties are more informative then a deterministic forecast. 93
94 The techniques used when performing weather forecasting, is often a combination of machine learning algorithms and statistical knowledge of weather phenomena. Unfortunately, the research has been extremely focused on weather predictions, so very few studies have applied the techniques used to other domains. In the future, we intend to use techniques from the weather forecasting domain for sales forecasting. In this study, however, we want to evaluate the techniques on publicly available standard regression problems. More specifically, this paper investigates how a technique originally proposed in the weather forecasting domain can be adapted to a wider range of regression problems. The key property of the technique, originally suggested in [TCM01] see section for details, is the ability to distinguish between forecast situations having higher or lower than expected. The main difference when moving outside the weather forecasting domain is that most regression problems lack the extensive statistical knowledge readily available for weather forecasting Method In weather forecasting, the boundaries of the bins are calculated statistically so the real value is equally probable to fall in any bin. This is only possible because weather data has been recorded in weather databases for many years. For most other domains, no such statistical databases exist, and it is therefore not possible to calculate a priori probabilities beforehand. Equal probability, Naive Bounds, Trimmed Naive Bounds and Genetic Algorithm which are presented below are potential alternatives which will be evaluated in the experiments. There are numerous ways to do probabilistic ranking and the three last techniques (Best, Variance and ANN) are straightforward alternatives to the binned ranking techniques. It must be noted that Best is only used for comparison, since it uses the real values to calculate the optimal result. Equal probability In the weather forecasting domain, the climatologic probability is calculated from historic data. Note that this data is more extensive then the data used to train the ensemble. For regression problems without extensive historical data the bin boundaries have to be estimated from the training data. Calculations based on a smaller subset (the training set) will of course be less reliable. Bins with equal probability is created using (35), i.e. by sorting the instances after the target variable and then selecting equally spaced instance values as boundary values. This will results in bins that are equally probable for the training instances. #instances b = sort ( f ( x) ); j = floor i * i # bins (35) 94
95 In the equation above bi is the boundary, #instances is the number of instances in the training set and #bins is the number of bins to calculate. Naive bounds The simplest way to calculate the bin boundaries is to divide the spread of the predictions equally among them. ( ) ( max ( E( x) ) min ( E( x) )) bi = min E( x) + * i (36) # bins Here E( x) is the ensemble prediction for the instances being ranked. The naive way to calculate the boundaries is sensitive to outliers since outliers could draw some of the bins to a range where very few predictions fall. Figure 25 shows an example of an ensemble with five members and their predictions. All instances have predicted values between 40 and 60, except two outliers. The outliers (instance 5 and 15) affect the calculation of the boundaries (the horizontal lines is the bin boundaries) in such a way that in reality only two of the bins are used for the ranking of the other instances M1 M2 M3 M4 M5 Figure 25 - Bins affected by outliers Trimmed naive bounds A value is normally defined as an outlier if it is more than two standard deviations from the mean. The sensitivity to outliers can be reduced by trimming the max and min using Formula(37). ( f x ) 4*std ( ) bi = mean ( f ( x) ) 2*std ( f ( x) ) + * i (37) # bins Figure 26 shows that the trimmed bins are a bit more suited to the problem than in Figure 25. In this example still only four of the bins are used, but it is an extreme example. 95
96 M1 M2 M3 M4 M5 Figure 26 - Trimmed naive bounds Genetic algorithms Another approach to choosing the bin boundaries is to optimize them on a training set where the goal is to minimize the error on the top 50%. In this study the optimization is done by using genetic algorithms. The bin boundaries are coded in a chromosome as a set of values (genes) representing the size of each bin. The chromosome is translated to the real bin boundaries by formula (38) 1 ( ) ( ( ) ( )) b = min f ( x) + max f ( x) min f ( x) * ( ( ) ( )) b = b + max f ( x) min f ( x) * 2.. n i i 1 n i= 1 gene n i= 1 gene i gene i 1 gene The initial population contains 300 randomly generated chromosomes. The fitness function is the mean absolute error (MAE) of the 50% instances with the highest bin scores. The population is evolved using the Genetic algorithm toolbox in Mat Lab 7.3. Standard settings are used for mutation, crossover and reproduction functions. The evolution is performed over 100 generations with roulette wheel selection. Mutations are done by adding a random number, or mutation, chosen from a Gaussian distribution, to each gene of the parent chromosome. The amount of mutation, which is proportional to the standard deviation of the distribution, decreases linearly during the evolution. The crossover function is a standard function where each crossover produces two new children from two parents. The crossover function is applied to 80% of all chromosomes at each generation except two elite (the fittest) chromosomes which are copied directly into the new population. The remaining vacant spots in the new generation are filled with chromosomes using standard reproduction. Variance The most straightforward way to rank the ensemble predictions according to error probability is to use the ensemble variance for each instance. The ensemble variance is the variance of the predictions done by the ensemble members. A 96 i (38)
97 lower variance indicates a more compact ensemble prediction and should correspond to a lower prediction error. To measure the performance of the technique using ensemble variance the mean absolute error (MAE) of the 50% instances with the lowest variance are calculated. ANN Here an ANN is trained to estimate a function describing the error made by the ensemble. This ANN is also trained on the training set, but here the individual predictions from the ensemble members are the inputs and the corresponding error made by the ensemble is the target. When the ANN is fully trained, it is applied to the test data, and the ANN output is noted for each instance. Finally the MAE is calculated on the ensemble error for the 50% instances for which the ANN has estimated the lowest ensemble error. Best The optimal result that could be achieved by any technique is calculated as the ensemble MAE for the 50% instances for which the ensemble had the lowest prediction error. It should be noted that this technique cheats by looking at the test instances and is only used for comparisons Experiments The study is evaluated on 10 data sets from the UCI repository. The ensembles used in the experiments consist of 20 ANNs. Each ANN is trained on an individual subset of the training set. The subset is created by randomly picking 80% of the training instances without replacement. The number of input variables is also reduced by selecting 80% of the features randomly. The purpose of this is, of course, to increase the diversity. Each ANN is initiated with random weights and with a random number of hidden units between 0 and 20. The training algorithm used to train the ANNs is Mat Lab s traingdx, which is a gradient descent back propagation algorithm, using a momentum term and adaptive learning rate. A validation set is used to decide when to stop the training. The validation sets is created by removing 20% of the training instances and assigning them to the validation set. Evaluation Mean absolute error (MAE) is used to measure the performance of the ensemble. The performance for each probability ranking technique on a certain data set is calculated as the MAE of the ensemble for the 50% of the instances that the technique ranked as having the lowest error. In the original study, see [TCM01], the technique was used to select the 10% worst and the 10% best instances, but here only the result for the top 50% instances are evaluated. This is required to 97
98 get a sufficient number of instances when using 10-fold cross validation as some datasets are quite small. As the data sets differ greatly in scale, MAE cannot be used to compare the techniques over all datasets. It is widely accepted that a unit free measure must be used if forecast methods are compared over several datasets [6]. To make the measure unit free, the percentage decrease in error (PDE) compared to the ensemble MAE is calculated. MAE( Ti, D j ) PDE ( Ti, D j ) = 1 (39) MAE, ( E D j ) Here, T is the technique and D is the data set and E is the ensemble Results All results presented in this section are the mean result over all ten folds, if it is not otherwise noted. In the sections below the result for the ensemble are first presented. Next the mode and the member support are presented for each technique and data set. Ensemble result Table 17 below presents the MSE for the ensemble on each data set. These results are only presented as a background to following sections as the results there shows how much this error could be reduced by the different techniques. Data MSE AUO BHD 2.38 IPD 0.10 MAC SER 0.55 WIS HUN 0.26 PHA SLE 2.30 VET Table 17 - Ensemble MSE Result using mode scores Table 17 shows the average PDE over all ten folds for each data set when the ensemble mode is used as scoring function. The PDE shows to which extent the ensemble error presented in Table 18 could be reduced by only choosing the top 50% ranked instances. 98
99 NAI TRI EQU GA VAR ANN BEST Dataset PDE R PDE R PDE R PDE R PDE R PDE R PDE AUO BHD IPD MAC SER WIS HUN PHA SLE VET MEAN ,7 Table 18 - PDE using mode scores (CD = 2.15) GA is the technique that achieves the highest PDE on average. It is 2 % better on average than the second best technique VAR, and 29% below the theoretical optimal result. All techniques except EQU perform reasonably well on average, but only GA, VAR and NAI achieves an error reduction on all datasets. NAI, EQU and ANN increase the error for at least one of the data sets. In the main experiments, the performance of the techniques is only measured on the top 50%. To show the general performance of the GA technique Figure 27 displays how the total error increases when choosing the instances in ascending order to their mode score. The solid describes the error when the instances are selected randomly. Note that the increase in % of total error decreases when the mode increases (the increase of error is correlated to the angle of the GA line). 100% 75% 50% 25% GA Mean 0% 0% 25% 50% 75% 100% Figure 27 - Gain chart for fold 1 of SER Results using member support 99
100 Table 19 shows the result when member support was used as scoring function. Note that the result for VAR, ANN and BEST is the same as in Table 18 since these techniques do not use the binned approach. All binned techniques here perform worse than VAR, and compared over all datasets, they are clearly worse than the result achieved by using the mode scoring function. EQU performance is not affected by support scoring but it is still far worse than the VAR technique. NAI TRI EQU GA VAR ANN BEST Dataset PDE R ACC R PDE R PDE R PDE R PDE R PDE AUO BHD IPD MAC SER WIS HUN PHA SLE VET MEAN Table 19 - PDE using support scores (CD = 2.15) Conclusions The most obvious result of this study is that the mode scoring function clearly outperforms the function using member support. The mode function reduces the error on the top 50% twice as much as when member support is used and it does it for all but one technique (EQU). The difference is also statistically significant for all techniques except EQU. Only GA, TRI and VAR succeed in reducing the ensemble error on the top 50% for all data sets. These techniques also have a higher performance in general compared to the other techniques. GA is the most promising of the techniques and is on average 2% better than the second best technique VAR, but this difference is not statistically significant. GA achieves an error reduction of 38% on average which has to be recognized as a good result. The optimal result achieved by BEST when it sorts the instances after the actual ensemble error is 67%. GA s result is a substantial decrease of the ensemble error and this information could be very useful for any decision maker. It is surprising that EQU, which is the technique most similar to the original technique proposed in [TCM01] performed by far the worst. The main reason for 100
101 this is probably that in the original study the bounds were calculated from extensive statistical knowledge which the data sets in this study lacked. The method used by EQU to produce equal likely bins could possible work for larger data sets, but the data sets used in this study were apparently too small to allow calculation of a priori probabilities. Even if EQU did not perform as well as expected, the GA version of the original technique did produce very promising results. This should be seen as a successful example of how techniques from a very mature domain closely related to information fusion, could contribute to fusion goals. Even if the domain does not match exactly, small adjustments can be done to techniques to make them compatible. Most important is the fact that the GA technique combined with the mode scoring function clearly outperformed the frequently used method of ranking by ensemble variance. 101
102 8.4 Inconsistency Friend or Foe This section is based on the paper inconsistency Friend or Foe [JKN07] Summary When evaluating rule extraction algorithms, one frequently used criterion is consistency; i.e. the algorithm must produce similar rules every time it is applied to the same problem. Rule extraction algorithms based on evolutionary algorithms are, however, inherently inconsistent, something that is regarded as their main drawback. In this paper, we argue that consistency is an overvalued criterion, and that inconsistency can even be beneficial in some situations. The study contains two experiments, both using publicly available datasets, where rules are extracted from neural network ensembles. In the first experiment, it is shown that it is normally possible to extract several different rule sets from an opaque model, all having high and similar accuracy. The implication is that consistency in that perspective is useless; why should one specific rule set be considered superior? Clearly, it should instead be regarded as an advantage to obtain several accurate and comprehensible descriptions of the relationship. In the second experiment, rule extraction is used for probability estimation. More specifically, an ensemble of extracted trees is used in order to obtain probability estimates. Here, it is exactly the inconsistency of the rule extraction algorithm that makes the suggested approach possible Background The most common rule extraction criteria are: accuracy (the extracted model must perform almost as well as the original on unseen data), comprehensibility (the extracted model should be easy to interpret by a human) and fidelity (the extracted model should perform similar to the original model). Two other very important criteria are scalability (the method should scale to networks with large input spaces and large numbers of weighted connections) and generality; i.e. the method should place few restrictions on network architectures and training regimes. Some authors add the criterion consistency; see e.g. [TS93]. A rule extraction algorithm is consistent if rules extracted for a specific problem are similar between runs; i.e. consistency measures how much extracted rule sets vary between different runs. The motivation is that if extracted rule sets differ greatly, it becomes hard to put a lot of faith in one specific rule set. Since G-REX (a rule extraction technique presented in section 6.1.2) uses GP to optimize the extracted representation, consecutive runs, even using the same opaque model, will normally result in slightly different extracted rules. In other words, G-REX is inherently inconsistent. This is also considered to be the main 102
103 drawback for G-REX (and other rule extractors based on evolutionary algorithms) in the survey reported by Huysmans, Baesens and Vanthienen; see [HBV06]. However in this paper, it is argued that the criterion consistency is insignificant. In addition, it is showed how the fact that a rule extraction algorithm is inconsistent can be utilized in order to perform probability estimation. Probability estimation trees Algorithms inducing decision trees are very popular, mostly because their ability to produce transparent and rather accurate models. In addition, decision trees work very well with large data sets and require a minimum of parameter tweaking. Although the normal operation of a decision tree is to predict a class label based on an input vector, decision trees can also be used to produce class membership probabilities; in which case they are referred to as probability estimation trees (PETs). The easiest way to obtain a class probability is to use the proportion of training instances corresponding to a specific class in each leaf. If 50 training instances reach a specific leaf, and 45 of those belong to class A, the assigned class probability for class A, in that leaf, would become 0.9. Consequently, a future test instance classified by that leaf would be classified as class A, with the probability estimator 0.9. Normally, however, a more refined technique, called Laplace estimate or Laplace correction is applied; see e.g. [PTR98]. The main reason is that the basic, maximum-likelihood estimate, described above does not consider the number of training instances reaching a specific leaf. Intuitively, a leaf containing many training instances is a better estimator of class membership probabilities. With this in mind, the Laplace estimator calculates the estimated according to equation p classa = k + 1 N + C (40) where k is the number of training instances belonging to class A, N is the total number of training instances reaching the leaf and C is the number of classes. In should be not that the Laplace estimator in fact introduces a prior uniform probability for each class; i.e. before any instances have reached the node (k=n=0), the probability for each class is 1/C. Provost and Domingo in [PD03] combine the Laplace estimator with the standard technique bagging [Bre96]. More specifically, an ensemble consisting of several, independently trained, PETs was created and final probability estimates were computed by averaging the individual probability estimates from all ensemble members. To obtain some diversity, each 103
104 PET was trained using only part of the training data, here a bootstrap sample. So, if the Laplace estimator for a certain leaf was: where P ( k x) N k k = K N + λ + k = 1 λ k (41) Nk the number of instances belonging to class k, N is the total number of instances reaching the leaf and λ k is the prior probability for class k, the ensemble probability estimates were calculated using below. L 1 = (42) P ( y x) P ( y x) L l = 1 In order to obtain what they call well-behaved PETs, Provost and Domingo changed the C4.5 algorithm by turning off both pruning and the collapsing mechanism. This obviously led to much larger trees, which, in fact, turned out to be much better PETs; for details see the original paper [PD03] Method The empirical study consists of two main experiments. The purpose of the first experiment is to investigate what level of consistency to expect from G-REX. In addition, G-REX accuracy is also compared to CART. The purpose of the second experiment is to evaluate whether G-REX can be used for probability estimation. The performance is again compared to CART. In both experiments, G-REX extracts rules from ANN ensembles. Each ensemble consists of 20 independently trained ANNs. All ANNs are fully connected feed-forward networks and a localist representation is used. Of the 20 ANNs, five have no hidden layer; ten have one hidden layer and the remaining five have two hidden layers. The exact number of units in each hidden layer is slightly randomized, but is based on the number of features and classes in the current data set. For an ANN with one hidden layer the number of hidden units is determined from (43) below. l ( ) h = 2* rand v * c (43) v is the number of input features and c is the number of classes. rand is a random number in the interval [0, 1]. For ANNs with two hidden layers, the number of units in the first and second hidden layer ( h 1 and h 2 ) are determined using (44) below. 104
105 ( v c) * ( v* c) h1 = + 4* rand * 2 c ( v* c) h2 = rand * + c c (44) Each ANN is, in an attempt to increase diversity, trained using only 80% (randomized without replacement) of available training instances. In this study, G-REX will be compared to CART, so the representation language is chosen to mimic CART; see the grammar presented using Backus- Naur form in Table 20 below. Rule ::= [If] [If] ::= If ([ROp]) then ([If R ]) else ([If R ]) [ROp] ::= [con] [> <] [C] [con] [> <] [con] [cat] == [C] [C] ::= Double in the range of associated [con] Class of associated [cat] [ R ] ::= Random class of target variable [con] ::= Continuous variable [cat] ::= Categorical variable Table 20 - G-REX Representation The GP settings used for G-REX in this study are given in Table 21 below. Number of generations 1000 Population size 1000 Crossover probability 0.8 Mutation probability 0.01 Creation type Ramped half and half Maximum creation depth 6 Length punishment (p) 0.01 Batch size 1 Table 21 - G-REX Settings Experiment 1 The purpose of Experiment 1A is to evaluate the consistency of G-REX. More specifically, the algorithm is run several times to extract a number of trees. The trees are then compared in order to determine whether they are similar or not. In 105
106 this experiment, similarity is measured only based on the predictions made by the different extracted trees. This means that the actual meaning of the rules is ignored. The evaluation is divided in two parts; intra-model and inter-model. Intra-model compares trees extracted from a mutual opaque model, while intermodel compares trees extracted from different opaque models, trained on identical data. The measure used is the proportion of instances classified identically by the two extracted trees. It should be noted that this measure is blind to how accurate the rules are, it measures only whether the two predictions are identical. In Experiment 1B, G-REX accuracy is compared to CART. The CART implementation used here is the one in Clementine. For the comparison, standard 10-fold cross validation is used. To obtain G-REX accuracy on a specific fold, ten G-REX trees are individually extracted from the ANN ensemble. Then the single tree having highest fitness is applied to the test set. This is a standard procedure when using G-REX, although the number of extracted trees to choose from is normally rather two or three than 10. But in order to investigate the consequences of G-REX inconsistency, all ten G-REX trees are also compared to the single CART tree. Experiment 2 The purpose of the second experiment is to investigate whether G-REX trees can be used for probability estimation. Here, three separate experiments are conducted. In Experiment 2A, G-REX and CART are again both run normally; i.e. just performing classification. In Experiment 2B, G-REX produces ten trees for each individual fold, and the tree with the highest fitness is applied to the test set. CART here uses all available data and consequently produces one tree per fold. Both G-REX and CART use the Laplace estimator to produce probability estimates based on the training instances. To determine the quality of the probability estimate, each tree ranks all test instances according to the estimates and discards the worst 20%. The accuracy is finally calculated for each tree on the remaining instances (Top80). In Experiment 2C, the G-REX trees are used as an ensemble with 10 members. The tree having the highest fitness is still used for prediction, but probability estimation is made by combining all ten trees; as described above. CART here first trains one tree normally; i.e. the parameter setting favors generalization accuracy. After that, ten additional CART trees are trained, prioritizing accuracy on the training data and using no pruning. These trees are consequently quite large. In order to introduce some diversity, each tree uses only 70% of the available training data. CART then uses the first tree for prediction, but the entire ensemble for probability estimation, again as described above. Using these techniques, it is very important to recognize that both G-REX 106
107 and CART still produce one, transparent, model. For the evaluation, the same procedure as in Experiment 2B is used; i.e. 80% of the test set instances are chosen for prediction, using the Laplace estimator. For all experiments, 10-fold cross validation is used. The CART implementation used here is the one in Mat Lab; i.e. the Tree Fit function in the statistics toolbox. It should also be noted that in this experiment, all datasets were stratified, which is the main reason for the higher accuracies overall Results Table 22 below shows the intra-model results for one fold from the Zoo problem. In this fold, the average consistency is for the training set and for the test set. The interpretation is that a pair of extracted trees, on average, agrees on 91.0% of all training set instances, and on 89.2% of all test instances. TRAIN R2 R3 R4 R5 Test R2 R3 R4 R5 R R R R R R R R Table 22 - Intra-model consistency (ZOO) Similarly, Table 23 shows the inter-model results for one Diabetes fold. Here the average consistency is for the training set and for the test set. TRAIN R2 R3 R4 R5 Test R2 R3 R4 R5 R R R R R R R R Table 23 -Inter-model consistency (PID) Table V below shows the overall results from Experiment 1A. On average, two extracted rules agree on approximately 89% of all instances. Somewhat surprising, this holds for both training and test sets, as well as intra-model and inter-model. Intra-model Inter-model Dataset TRAIN TEST TRAIN TEST BLD CHD PID SMR WBC
108 ZOO MEAN Table 24 - Average consistency over all pairs and folds Table 25 below shows the comparison between CART and G-REX in Experiment 1B. The first two columns contain the overall test set accuracies obtained using 10-fold cross validation. The last three columns compare all extracted G-REX trees to CART. The numbers tabulated are the number of wins, ties and losses for G-REX, when comparing the ten extracted trees on every fold to the corresponding CART tree. The ensemble clearly outperforms CART but the results of the ensemble are not presented as the focus lies on comparing G- REX and CART. Dataset ACC CART ACC G- REX #Wins G- REX #Ties #Wins CART BLD CHD GCD JHI LYD PID SMR WBC WIN ZOO MEAN Table 25 - Accuracy comparison between G-REX and cart Although it interesting to note that G-REX outperforms CART on 8 of the 10 data sets, this is not the major point here. The most important observation is instead that a huge majority of all individual G-REX trees are at least as accurate as their CART counterpart. Only 13.2% of all GREX trees have lower accuracy compared to the corresponding CART tree. Table 26 below shows the results from Experiment
109 Dataset Normal Laplace Top 80 Ens. Top 80 CART G-REX CART G-REX CART G-REX BLD CHD GCD JHI LYD PID SMR WBC WIN ZOO MEAN Table 26 - Comparison between methods There are several interesting results in Table VII. First of all, it should be noted that a single G-REX tree most often outperformed the corresponding CART tree; both in Experiment 2A and 2B. In addition, it is obvious that both techniques could utilize the probability estimates obtained in order to rank instances based on how hard they are to predict. Arguably, the most interesting observation is, however, the fact that G-REX using the ensemble approach obtained a clearly superior probability estimate Conclusions Consistency is a hard criterion to assess. Although it must be considered important to have at least fairly high consistency, it is not obvious exactly what level this translates to. As comparison; tree inducers like CART will produce identical trees every time, so their consistency would be 100%. With this in mind, the consistency obtained by G-REX initially raised some questions. As seen in Experiment 1, on some data sets, more than 15% of all instances are, on average, classified differently by two GREX runs. Is it really possible to put faith in one specific extracted rule when another run might produce a different rule, disagreeing on 15% of all instances? The answer to this question is not obvious. Intuitively, most data miners would probably want one accurate and comprehensible model. On the other hand, most individual G-REX runs do produce exactly that; i.e. a compact yet accurate tree. As shown in Experiment 1, any single G-REX tree is most often more accurate than the corresponding tree induced by CART. Apparently there are several descriptions (models) of the 109
110 relationship that all have similar accuracy. Using a tree inducer you get one specific model, but why is that better than obtaining several, slightly different, models, each having high accuracy? Based on this, it can be argued, that for rule extraction algorithms, low consistency should not be considered a major problem. Regarding the experiment with probability estimation, it is obvious that G-REX is more than capable of producing good probability estimates. Especially the results where G-REX used all extracted trees in order to obtain a probability estimate are promising. Clearly, it is the inherent inconsistency of G-REX that makes this approach possible, without having to use separate parts of the data set for each ensemble member. Considering the relatively small amount of extra work needed to extract several trees instead of one, this is an encouraging result for both G-REX and rule extraction in general Discussion and future work The ultimate goal of rule extraction is of course one accurate and comprehensible model. At the same time, it is well-known that sufficiently diverse ensembles are more accurate than single models. One interesting question is therefore whether extracted trees could be fused together to increase the performance. Another option is to keep all trees, and somehow try to visualize them together in order to increase the understanding of the underlying relationship. In this study, G-REX extracted rules from an ANN ensemble. It would be interesting to investigate whether probability estimates based on the ensemble could be used in some way to improve the probability estimation performed by the G-REX tree. G-REX has previously been used to produce different types of regression trees. With this in mind, one obvious study is to try to utilize G-REX s inconsistency, in order to produce probability estimates for regression problems such as time series. This should be compared to the standard way of producing probability estimates for times series; i.e. to post-process the output from each ensemble member statistically (Ensemble-MOS) Considerations Most data sets in the experiments have a skewed distribution (last column in Table 6) but no correction for this is done in the experiments. It is possible that a correction towards the true prior probabilities could further improve the results. 110
111 8.5 Genetic programming A Tool for flexible Rule Extraction This section is based on the paper Genetic programming A Tool for flexible Rule Extraction [KJN07] Summary We have previously suggested a rule extraction algorithm named G-REX, which is based on genetic programming. One key property of G-REX, due to the use of genetic programming, is the possibility to use different representation languages. In this study we apply G-REX to estimation tasks. More specifically, three representation languages are evaluated using eight publicly available datasets. The quality of the extracted rules is compared to two standard techniques producing comprehensible models; multiple linear regression and the decision tree algorithm C&RT. The results show that G-REX outperforms the standard techniques, but that the choice of representation language is important. This is not a problem for G-REX, since it can use any representation language. In addition, it is easy to perform initial experiments in order to choose the most suitable. The study also shows that it is often beneficial to guide the GP search in some way. In the most successful technique LINREG, GP was used to divide the dataset in subsets on which standard linear regression was applied. All representations could actually be used to construct the rules produced with LINREG but the GP search seldom reached these solutions due to the enormous search space Background In the only previous estimation study [JKN04], G-REX used a simple representation language described using Backus-Naur in Table 27. Furthermore, the use of G-REX for estimation tasks, was only tested in the market response modeling domain. With this in mind, the overall purpose of this study is to evaluate G-REX using two new representation languages for estimation Method This study evaluates if GP, and more specifically G-REX, can be used to extract accurate comprehensible rules from an ensemble of Artificial Neural Networks (ANN). The extracted rules are evaluated against the tree algorithm C&RT, and multiple linear regression. The following sections will describe the experimentation. Ensemble Creation The ensembles used in the experiments consist of 20 fully-connected ANNs, each having one hidden layer. Each ANN is trained on an individual subset of the training set. The subset is created by randomly picking 80% of the training instances without replacement. The number of input variables is also reduced by 111
112 selecting 80% of the features randomly. The purpose of this is, of course, to increase the ensemble diversity. Each ANN is initiated with random weights, and with a random number of hidden units which is calculated by (45), where R rand is a random number between 0 and 1. ( ) # hidden = floor R *20 (45) The training algorithm used to train the ANNs is Mat Lab s traingdx, which is a gradient descent back propagation algorithm, using a momentum term and adaptive learning rate. A validation set is used to decide when to stop the training. The validation sets is created by removing 20% of the training instances and assigning them to the validation set. 112
113 Representation languages As described above, one of G-REX main strengths is that it can use many different representation languages. To evaluate the importance of representation language, the experiments are performed evaluating three different representation languages. The representation languages used by G-REX are presented using Backus-Naur in the tables below. Rule ::= [If] [If] ::= If ([cond]) then ([exp]) else ([exp]) [cond] ::= [con] [func] [C] [con] [func] [con] [cat] = [C] [exp ] ::= [If ] [R ] [func] ::= > < [C] ::= Double in the range of associated [con] Class of associated [cat] [ R ] ::= Double in range of target variable [con] ::= Continuous variable [cat] ::= Categorical variable Table 27 - BNF for normal representation (NORM) NORM is the simplest representation language for a regression tree and uses the same set of functions and terminals as C&RT. The only difference is that C&RT s syntax only allows comparisons between a variable and a constant value in a condition, while NORM also allows comparisons between two variables. NORM is the original representation language used for G-REX estimation in [JKN04]. If ( X1 > 5 ) If ( X2 < 3 ) If ( X1 > X2 ) If ( X3 = 1 ) Figure 28 - Example rule for NORM The if-statements can be seen as a series of conditions dividing the instances of a data set into subgroups depending on the value of some input variables. Each subgroup is associated with the value of a certain terminal. For NORM the terminal is always a constant which means that all instances (known and novel) belonging to a certain terminal will be predicted to have the same value. The conditions of the if-statements and the terminal constants are of course optimized, using GP, on the training data. 113
114 Rule ::= [If] [If] ::= If ([cond]) then ([exp]) else ([exp]) [cond] ::= [con] [func] [C] [con] [func] [con] [cat] = [C] [exp ] ::= [If ] [linreg] [func] ::= > < [C] ::= Double in the range of associated [con] Class of associated [cat] [linreg] ::= [con] * [linc] + [linc] [linc] ::= Double value calculated using the least square method on the training instance reaching the associated [linreg] [con] ::= Continuous variable [cat] ::= Categorical variable Table 28 - Representation with linear regressions (LINREG) The LINREG representation extends the NORM by allowing linear regression as terminals. Using this grammar, the representation language becomes more powerful, since a terminal can predict different values for each of the instances it is associated with. The predicted value pi for instance i is calculated using (46) below; where z i is the value of one of the input variables. K and M are constants found by the linear regression. p = K * z + M (46) i When a linear regression expression is created K and M are calculated by the least square method [Dun03]. The calculations are done on the training instances that are associated with the linear regression terminal when the rule is first created. During evolution, mutation and crossover are only applied to the if-statements and its conditions, which will affect how the instances are associated with the terminals but not the linear regressions which will remain the same. Mutation can introduce new linear regression terminals during the evolution. Figure 29 below shows an example of a rule created following LINREG. i 114
115 If ( X1 > 5 ) If ( X2 < 3 ) If ( X1 > X2 ) X1 * X1 * If ( X3 = 1 ) X2 * X2 * X1 * Figure 29 - Example rule for LINREG The last representation FREE presented below is similar to LINREG with two exceptions. First, each subgroup of instances can be associated with a combination of several linear regressions, making it possible to base the prediction on several input variables. Secondly the linear regressions are created randomly and are affected by the genetic operations in normal fashion. 115
116 Rule ::= [If] [If] ::= If ([cond]) then ([exp]) else ([exp]) [cond] ::= [con] [func] [C] [con] [func] [con] [cat] = [C] [exp ] ::= [If ] [+] [*] [func] ::= > < [C] ::= Double in the range of associated [con] Class of associated [cat] [+] ::= [term] + [term] [*] ::= [con] * [C] [term] ::= [+] [*] [C] [con] ::= Continuous variable [cat] ::= Categorical variable Table 29 - Free representation (FREE) The linear regressions are combined by addition, in the same way as in multiple linear regressions. This is in essence a way to evolve different multiple linear regressions for different parts of the datasets. If ( X1 > 5 ) If ( X2 < 3 ) + * + * * X X2 0.2 X1 1.3 * * X2 0.4 X1 2.2 Figure 30 - Example rule for FREE Results In previous studies when G-REX has been used to extract regression rules, Equation (26) has been used to calculate the fitness of a rule. Here the factor p is used to balance accuracy versus comprehensibility, but it will have different effect depending on the magnitude of the error. Before G-REX is applied to a new dataset, trial experiments have to be performed to decide an appropriate size of the factor. To achieve a more true balancing the fitness function below is used. k = n ( E( ik ) Rr ( ik ) ) k = 1 Regfitnessr = 1 sizer * p C (47) 116
117 C is initially set to the sum of the target variable and is then updated after each generation according to C*1.5 if Regfitnessbest > 0.9 C= C if 0.1 Regfitnessbest 0.9 C *0.5 if Regfitnessbest < 0.1 (48) C is used to normalize the ae in a way that makes the balancing of the length of the rule independent of the magnitude of the error. Evaluation To evaluate the benefit of using rule extraction, G-REX is evaluated against algorithms that produce comprehensible models. In this study G-REX is evaluated against multi linear regression (MREG) and C&RT [BFO+83] a common decision tree algorithm. Both techniques produce transparent models, normally regarded as comprehensible. The experiments were performed on eight datasets from the UCI repository using standard 10-fold cross validation with stratification to ensure that each fold was representative. Details regarding the datasets used in this study can be found in section This study focus on estimation problems where the error measure GMRAE recommended in section cannot be applied as it is relative to random walk. But the GMRAE can still be used by simply exchanging the reference to the rw in (16) with the mean of the currently known actual values as done in (49). RAE s = F F m, s mean A s A s (49) Results The result presented in Table 30 is the average GMRAE of the ten folds of each dataset. ENS is the result for the ensemble which is the target for the G-REX. The bold results are the best result for each dataset excluding the ensemble. The results are relative to the mean prediction, so a value less than one, means that the results are better than the mean prediction. ENS NORM FREE LINREG MREG C&RT AUT DIA BHD MAC PHA SLE
118 VET WIS MEAN Table 30 - GMRAE on individual datasets Conclusions The results show that G-REX outperforms the standard regression techniques, but that the choice of representation language is important. This is not a problem for G-REX, since it can use any representation language. The study also shows that it is often beneficial to guide the GP search in some way Considerations MREG and C&RT both have quite simple representations while G-REX s FREE and LINREG are slightly more complex. Hence the comparison is not totally just but since G-REX produce very short rules they are still deemed to be at least as comprehensible as the simple representations. G-REX could of course have been compared against techniques as Cubist that produce more complex expressions but these rules are often too complex to be considered as comprehensible. 118
119 8.6 ICA Summary Neural networks have many successful applications for market response modeling. However the downside of neural networks is the many settings that is required and that the model is incomprehensible. The aim of this study is to show how these two deficiencies can be handled by rules extracted from an ensemble of opaque models. Experiments on a dataset of weekly sales for nine store specific articles show that an ensemble of different models outperforms several standard techniques producing comprehensible models. Further experiments also show that it is more favorable to extract a comprehensible model from the ensemble than to create one from the original dataset Background Retailers have a need for accurate market response models as an accurate model facilitates an optimization of stock levels and could give improved decision support to campaigns planners. A market response model is supposed to capture how the casual variable such as price, marketing efforts and previous sales affect future sales. ICA Sverige AB Together with its retailers ICA Sverige AB is Sweden s largest food retail company with more than 50% of the market ICA had 1382 divided into four concept categories: ICA Nära comprises smaller, conveniently located food stores that provide good service, a streamlined product range and high-quality fresh foods. ICA Supermarket s stores are conveniently located in the areas in which our customers live or work, and they are often the main place used by our customers to buy most of their groceries. ICA Kvantum comprises larger supermarkets offering everyday food products, special foods for allergy sufferers, eco-labeled products and specialty items from near and far. ICA Kvantum also offers a large number of products in the beauty, health and media ranges. Maxi ICA Stormarknad comprises hypermarkets where customers will find everything they need in a single location at advantageous prices. The stores have generous opening hours and are conveniently located for our car-borne customers. 119
120 Major marketing efforts such as TV-commercials and national- or concept wide campaigns are done centrally by ICA Sverige AB but a retailer can also do local marketing efforts. ICA makes all purchases centrally and the retailers make their individual purchase from ICAs central organization. Hence an accurate market response model is needed for each store, to facilitate optimization of central purchases and for giving purchasing recommendations to each store. Retailers can always decide exactly how much they want to buy but since they are not always aware of central marketing efforts and how they affect sales recommendations are needed. All sales are recorded locally in each store and then collected by the central organization and stored for three years. Central marketing efforts are also recorded and stored for the same period of time. In an initial attempt to evaluate different market response modeling techniques 40 articles were selected from four stores (one of each concept). Each article was selected to represent a typical sale pattern such as campaign- event- weather- price- season- or holiday sensitivity. Only a few of these articles could be modeled with a reasonable performance since they had a rather low sale frequency. It would probably be better to aggregate the sales for each store concept to increase the frequency. Another problem was that local marketing efforts made by the retailers was not recorded and hence added much noise to the data. Market response modeling Traditionally market response modeling have been done using econometric techniques which rely on statistical techniques to estimate relationships of models specified on the basis of theory, prior studies and domain knowledge [ARM01]. However, in the latest decade researcher in market response modeling and forecasting have shown a great increase in interest for artificial neural networks as modeling technique. Several studied such have shown that neural networks most often are superior to econometrics models; for a survey see [Zha04]. The main reason for the superiority of neural networks is thought to be that nonlinearity is often present in market response data. A certain marketing effort will for example always affect the sales positively but each additional unit of marketing effort will result in less incremental sales than the previous. Other nonlinearities that a sales response function can show includes minimum and maximum sales potential, saturation, threshold, and/or marketing mix interactions effects. Nonlinearities like these are hard to formalize in econometric models but could easily be modeled by a neural network as they are general function approximators. Another advantage is that the relationship is not needed 120
121 to be known in advanced as long as there exists sufficient data that capture phenomena. Artificial neural networks for market response modeling Neural networks (described in section 5.3) is a powerful technique but also requires that several model parameters are set in an appropriate way to achieve optimal results. These settings are also problem dependent making the problem even more cumbersome as no single setting can be recommended for all problems or even two datasets with the same input variables. When designing a neural network model important design choices includes number of hidden layers, number of units in each layer and which activation functions that should be used. Training a neural network also includes choices of training function and when to stop the training. When performing market response modeling and other forecasting application [ZPH98] gives a set of recommendations based on a survey of most promising results in the field. One hidden layer is often enough but two hidden layers may give better results for some problems. Two hidden layers also results in a more compact architecture which achieves a higher efficiency in the training process. The number of units needs to be selected using initial experiments as none of the existing heuristics works for all problems. As forecasting problems involves learning about deviations from the average the hyperbolic activation function most suitable for the hidden units while the output units should have linear activation functions. BFGS or Levenber-Marquardt suggested as training function as these have faster convergence, are more robust and that their parameters can be set automatically. Using these recommendations the only important decision that remains is the number of units that should be used in each of the hidden layers. One approach to handle this decision is to create an ensemble (see section 5.5) of neural networks where the numbers of hidden units are chosen randomly. For a single network choosing the number of hidden units randomly would be an extremely bad idée as the performance of a single network can vary greatly depending on the number of hidden units. Actually as a neural network starts with randomly assigned weights the performance of a single neural network can vary only depending on the initial weights making it hard to predict it performance 121
122 beforehand. However for an ensemble a diversity of its members prediction is actually favorable as they hopefully will cancel each other s error [KV95]. Rule extraction Ensembles and neural networks are inherently opaque models as they are hard or impossible to interpret for a decision maker. This is a serious problem when modeling market response as there are many factors such as competitor action or weather that are uncontrollable. If an unforeseen event would occur it is crucial that a decision maker can understand how this will affect the current forecast. A comprehensible model can also give insight to which variables that drives the sales or how to best use a certain marketing channel. One common approach to this problem is called rule extraction (see section 6.1), i.e. the process of extracting a comprehensible model from the opaque models while retaining a high performance. Several studies, for example see [JKN03, KJN07, KJN08] has shown that it is most often favorable to extract a model than to create a comprehensible model from the original data Method To show how rule extraction and ensembles could be used for market response modeling two experiments will be performed sequentially. First an ensemble will be created and evaluated against three different transparent techniques and then the ensemble will be used for rule extraction and the extracted rules will again be compared to the other transparent models. Ensemble creation As neural networks often have been applied successfully to market response modeling they will be used as the foundation of the ensemble. The creation of each individual network will follow [ZPH98] recommendations and use two hidden layers where the units have hyperbolic activation function and one output layer with a single output unit with a linear activation function. Each layer will be created with a random number of hidden units ranging from 2 to 20 units in hope that this will increase the diversity of the networks while it in the same time removes a critical design choice. Even if the initial study showed that an neural network ensemble were superior to any single technique it also showed that the ensemble accuracy could be improved by including other model types. Thus a mixed ensemble will also been used in this study as the highest possible accuracy is required but also to show the generality of the rule extraction process. After the neural network ensemble is constructed it will be seen as single model with which outputs the average prediction of all networks. Next two new models 122
123 will be created using multiple linear regression and Mat Labs Tree fit. Finally the predictions from all three models will be averaged to give a single output. Rule extraction The final ensemble will be used to create a new dataset for the rule extraction process by replacing each target value in the training set with the ensemble prediction. G-REX will then be applied to each dataset ten times and the average result will be reported. For this study G-REX will optimize the MAUPE in its fitness function as this facilitates an unbiased optimization and that all G-REX settings can be the same for all datasets. n y i yi i= 1 ( yi + yi ) / 2 MUAPEfitness = 1 size* p n (50) Number of generations 100 Population size 1000 Crossover probability 0.8 Mutation probability 0.01 Creation type Ramped half and half Maximum creation depth 6 Length punishment (p) Batch size 1 Table 31 - G-REX settings 123
124 Evaluation As the aim of G-REX is to produce a transparent comprehensible model with high precision it will be evaluated against two transparent techniques, multiple linear regression (MREG) and Mat Labs Tree Fit (A decision tree technique similar to C&RT). MREG and Tree Fit are based on two quite different representations and will therefore be evaluated against two different G-REX representations. MREG G-REX will compared to LINREG (described in Fel! Hittar inte referenskälla.) and Tree FIT to NORM (Described in Figure 28) as these are deemed to be the most comparable G-REX representations. As G-REX is a nondeterministic technique and its results can vary slightly for each run it will be executed 10 times for each dataset and the average result will then be reported. Each model will be evaluated against MUAPE GMRAE and R 2 as each metrics measures different properties of the models. The Wilcoxon test will be used to test for significance. Datasets In this study a subset of the most frequent selling articles was chosen to be able to demonstrate the use of ensembles and rule extraction for market response modeling. The following nine articles were selected from four specific stores: Store Concept Article Sales pattern Supermarket Milk Regular Supermarket Cream Regular/Holiday/season Frequently marketed snack. Supermarket Risifrutti Maxi Potato chips Event sensitive Maxi Sour cream Holiday sensitive Maxi Ice cream Weather sensitive Maxi Butter Frequently marketed Kvantum Butter Frequently marketed Nära Butter Frequently marketed Table 32 - Selected articles The dataset contained between 80 to 105 weeks and were divided by assigning the first 80% as training set and the last 20% as test set. Each dataset contained the variable presented in the table bellow but some datasets did not have any campaign efforts recorded. Variable Tap delays Weekly sales 5 TV commercial 3 + Predicted Week 124
125 Loyalty card 3 + Predicted Week National campaign 3 + Predicted Week Concept campaign 3 + Predicted Week Discounted 3 + Predicted Week Salary week 3 + Predicted Week Table 33 Dataset variables As the information about local campaigns was missing outliers that occurred without any central campaign being present was replaced with the average of all sales. 125
126 8.6.4 Results Each model is evaluated against GMRAE, MUAPE, RMSE, R 2, and MAD and a Wilcoxon test is performed for each metric to test for significance. The mean is not presented for RMSE and MAD as these metrics are not unit free and therefore cannot be average in a fare way. A single ensemble is trained for each dataset and then used as target for the rule extraction and for the comparison against both transparent techniques. Table 34 shows the results for the mixed ensemble, it is these results that are used when the multiple linear regression- and Tree Fit models are evaluated in Table 34 and Table 35. A grey cell in these tables signifies that the model is better than the ensemble and the Wilcoxon test are based on the differences of the ranked difference between the ensemble and the evaluated model. Concept Article Mixed Ensemble GMRAE MUAPE R2 Super Milk 77.3% 6.00% Super Cream 70.5% 19.9% Super Risifrutti 57.2% 23.0% Kvantum Butter 40.2% 26.2% Maxi Chips 56.9% 23.9% Maxi Sour cream 85.2% 65.8% Maxi Butter 98.8% 14.1% Maxi Icecream 47.0% 35.6% Nära Butter 78.7% 12.6% Mean 68.0% 25.2% Table 34 - Mixed Ensemble Results The mixed ensemble is better than the random walk (predicting that next week sales will be the same as the last week) for all datasets as the GMRAE is below 100%. However the unbiased percentage error ranges for a low 6% for milk to a rather high 65% for sour cream and all predictions have also have a low R 2 value. Concept Article Multiple Linear Regression GMRAE MUAPE R2 Super Milk 94.6% 6.10% Super Cream 97.7% 22.7% Super Risifrutti 63.9% 23.3% Kvantum Butter 72.9% 34.1% Maxi Chips 58.0% 21.6% Maxi Sour cream 108.1% 71.9%
127 Maxi Butter 157.9% 17.9% Maxi Icecream 62.1% 37.6% Nära Butter 83.9% 14.4% Mean 88.8% 27.7% Wilcoxon test Table 35 Multiple Linear Regression vs Ensemble (CV=6) Table 35 above presents the results for a multiple linear regression model optimized using least squares. When compared to the ensemble the regression model has better R 2 for three of the datasets and actually have a higher average R 2, however the correlations are still very low and the differences is not significant according to a Wilcoxon test where the critical value is six and the test value for R 2 is 14. When compared to the ensemble using MUAPE and GMRAE the regression model is outperformed on all datasets and this is of course a significant difference. Table 36 below shoes the results of the rules extracted from the ensembles using G-REX and the LINREG representation. Concept Article G-REX Linear Regression GMRAE MUAPE R2 Super Milk 86.8% 6.20% Super Cream 78.4% 21.1% Super Risifrutti 60.1% 22.7% Kvantum Butter 66.9% 33.0% Maxi Chips 45.2% 19.6% Maxi Sour cream 95.9% 71.4% Maxi Butter 140.1% 14.7% Maxi Icecream 52.9% 33.2% Nära Butter 96.7% 14.0% Mean 80.3% 26.2% Wilcoxon test Table 36 - G-REX Linear vs Multiple Linear Regression (CV=6) Most extracted rules have worse performance the ensemble and for some dataset the difference are quite large. However the extracted rules keep enough of the ensembles superior performance to outperform MREG significantly on both GMRAE and MAUPE. Concept Article Tree Fit GMRAE MUAPE R2 Super Milk 102.8% 6.90% Super Cream 78.3% 25.5%
128 Super Risifrutti 95.9% 30.3% Kvantum Butter 45.2% 32.7% Maxi Chips 69.2% 30.5% Maxi Sour cream 98.6% 27.9% Maxi Butter 90.8% 12.8% Maxi Icecream 55.9% 38.8% Nära Butter 107.8% 17.1% Mean 82.7% 24.7% Wilcoxon test Table 37 Tree Fit vs Ensemble Compared to the ensemble the Tree Fit models are significantly worse regarding GMRAE. The ensemble is also clearly superior regarding MUAPE with seven wins and to losses. According to a Wilcoxon test this difference is not significant as there is a big difference in MUAPE (in favor for Tree Fit) for the sour cream dataset. Rules extracted with G-REX Naïve regression BNF (presented in Table 38) have a representation similar to Tree Fit. Concept Article G-REX Naive Regression GMRAE MUAPE R2 Super Milk 107.2% 6.70% Super Cream 69.5% 18.3% Super Risifrutti 66.3% 24.4% Kvantum Butter 55.6% 30.3% Maxi Chips 52.1% 22.7% Maxi Sour cream 140.6% 75.7% Maxi Butter 142.4% 16.3% Maxi Icecream 46.7% 31.9% Nära Butter 94.1% 13.9% Mean 86.1% 26.7% Wilcoxon test Table 38 G-REX Naive Regression vs. Tree Fit In this case there are no direct significances in the results. Nevertheless the G- REX Naïve regression rules are superior in regards of MUAPE and only loses against Tree Fit on the same dataset where Tree Fit also outperformed the ensemble. If these two datasets is disregarded the G-REX rules are significant as it outperforms Tree Fit on all other datasets Discussion 128
129 The datasets chosen for this study are the ones with the highest sale volumes as these tend to be a bit easier to predict. A MUAPE of around 15% could be handled by a keeping a safety stock but this level is only achieved for less than half of the datasets. In the other cases the prediction would have to be adjusted by hand and thus would require comprehensible models to facilitate more rational adjustments. When comparing MREG with the G-REX Linreg representation, Linreg is better on all datasets where the ensemble outperformed MREG. The extracted rules are also quit compact and most often only consist of a combination of a few linear regression. G-REX naive regression representation achieved similar results as Tree Fit for both GMRAE and R 2. This is probably due to the fact that both representations only guess a few constant values close to the average and thus do not follow the weekly changes closely. However G-REX superior result on MUAPE shows that these constants can be approximated in a better way using the ensemble output. Both Naïve regression and Linreg is slightly more powerful representations than MREG and Tree Fit as they allow splits based on comparison between variables. G-REX uses this feature fairly often to signify trends between the sales of two weeks. Two common splits are when the current week is compared to the previous week and when the current week is compared to the same week in the previous month. Figure 31 shows a typical rule for the cream dataset where the fist split is a comparison of the current week (fsg0) and the same week last month (fsg-4). If ( fsg0 > fsg-4 ) Fsg-1*0,19+222,3 If (Salary week) Fsg-1 Fsg-0*0,93+68,3 Figure 31 - Example Linreg rule for the Cream dataset This type of comparison is natural for time series and does not add much complexity for a decision maker. However this could of course be one reason to G-REX superior performance as neither Tree Fit nor MREG can do similar comparisons. On the other hand this is also G-REX strength as it allows representation language to be tailored to the problem at hand. All models have have very low R 2 values which is a sign that the underling patterns are extremely distorted. As the sale volumes are low the underlying patterns tends to disappear if the slightest noise is present. All G-REX models is quite small which is a good sign as a more complex rule surely would be over 129
130 fitted and thus be based on the noise instead of the real underlying pattern. Sales should probably instead be aggregated for all stores or all stores of the same concept. The prediction for the aggregated sales could then be broken down according to each stores relative proportion of the total sale Conclusions Even if the chosen datasets are the ones with the highest sale volumes the level is still too low to facilitate prediction with a sufficient precision for a fully automated DSS. In the cases that the models do succeed to achieve a god enough MUAPE it is done with predictions close to the average. This can also be seen in the fact that all models have a very low R 2 value which means that they describe a very small part of the variations of the sale pattern. A better approach is probably to pool the sales for several stores and make predictions for the aggregate sales. These predictions could then be broken down to individual stores according to the stores relative proportion of the aggregate sales. Even if the ensemble does not achieve sufficient precision G-REX succeeds in extracting short rules with significantly lower MUAPE and GMRAE than multiple linear regression. Compared to Tree Fit, G-REX is also significantly better in MUAPE when evaluated on the datasets where the ensemble outperforms Tree Fit (7 of 9). It should be noted that the representations used by G-REX are slightly more powerful than MREG and Tree Fit as it allows comparison between variables and thus roughly can model trends. On the other hand this is also G-REX strength as it allows representation language to be tailored to the problem at hand. 130
131 9 Conclusions In this section the more general conclusion that can be made based on the presented theory cases studies will be presented. The conclusion will be presented in accordance to the related objective. 1. Evaluate usability issues for predictive modeling in situations which suffers from imprecision, uncertainty and/or lack of relevant data. Ensemble techniques should be a natural choice for most decision support since they have been proven to be superior compared to other single techniques when compared over a large set of problems. This is also apparent the experiments of 8.1, 8.4, 8.5 and 8.6 where the ensembles outperformed the other techniques. However the superiority of ensembles does not mean that they always achieve a sufficient performance to facilitate an automatic system. Study 8.6 clearly show this as the ensemble here were significantly better than the other techniques but still failed to achieve of adequate performance for several data sets. Most often the reason for inadequate performance is that the data is imprecise, contains large uncertainties or that important variables simply are missing. Again study 8.6 is a good example, first the data concerning the campaigns were very imprecise as they did not contain any information of the cost or impact of a campaign and secondly important variables describing local campaign were missing. If the ensemble cannot make a good enough prediction it has to be adjusted manually in some way. However as an ensemble most often is opaque i.e. its members are opaque; it is hard or impossible to interpret, which rules out rational adjustments. Apparently ensembles alone are not a good basis for a decision support system that needs to work in an environment with imprecise uncertain data. Rule extraction has often been used to solve the problem of making opaque models transparent and comprehensible and is thus motivated by these results and all cases studies show that G-REX is well suited for this task. 2. Survey the literature for rule extraction techniques suitable for extracting comprehensible models from ensembles. Only pedagogical techniques should be considered when extracting rules from ensembles as they are the only kind that is general enough to deal with an ensemble of arbitrary members. G-REX has previously been shown to outperform other pedagogical techniques in regards of accuracy and especially in regards of comprehensibility and is thus a stronger candidate when choosing rule extraction technique. 131
132 3. Evaluate the need of improvements for the rule extraction technique G-REX. Increasing comprehensibility When compared with other techniques G-REX ability to create short and comprehensible rules has been strongly emphasized. In general G-REX performs better or comparable to transparent techniques but does this with much shorter and comprehensible rules. Yet when closer inspected these rules sometimes contains unnecessary elements (introns) and could therefore be made even more compact. A more subtle issue is, however, the fact that rules often include perfectly logical and semantically correct parts that still do not affect the predictions from the extracted model. To retain G-REX strength of being independent of its representation the simplification also has to be done in a general way. A novel approach to this task was successfully presented and evaluated in 8.1. The method can be seen another application of rule extraction. Instead of the ensemble, the extracted rule is used as the target for another rule extraction process. The technique performs remarkably well and succeeds in simplifying the extracted rules for 13 of the 16 data set used in the study. It should be noted that the rules that were simplified were already are significantly shorter than the once produced with the transparent models and the simplified rules is significantly shorter than the extracted rules. Evaluating G-REX against the consistency criteria. It is neither straightforward nor trivial to evaluate consistency especially in terms of setting a level for an acceptable consistency. Tree inducers are deterministic and will always produce the same tree which translates to a consistency of 100%. As seen in Experiment 1 of study 8.4 G-REX obtains a consistency as low as 85% for some data sets. This leads to the question of which rule that should be used if they have the same accuracy but differs on 15% of the predictions. Intuitively, most data miners would probably want one accurate and comprehensible model. On the other hand, most individual G-REX runs do produce exactly that; i.e. a compact yet accurate tree. As shown in Experiment 1, any single G-REX tree is most often more accurate than the corresponding tree induced by CART. Apparently there are several possible models of the relationship that all have similar accuracy but there is no way to judge which the true one is based on the data. A tree inducer would produce a single model but why would that be the true one? Techniques producing a single model would instead risk making the 132
133 decision maker over confident that the produce model is the only true possibility. Inconsistency could instead be regarded as an advantage as it could be used as another measure of the belief a decision maker should put in models from a certain data set. Hence for rule extraction algorithms, low consistency should not be considered a major problem per se. As seen in the second experiment another advantage of slightly inconsistent accurate models is that they could be used to produce improved probability estimation. Considering the relatively small amount of extra work needed to extract several trees instead of one, this is an encouraging result for both G-REX and rule extraction in general. Evaluating G-REX for estimation tasks. Previous work on G-REX has mainly been focused on classification. Only a single study demonstrating the concept has been performed for estimation tasks. However both study 8.5 and 8.6 clearly shows that G- REX is well suited to also handle both estimation tasks and time series prediction. G-REX GP foundation also facilitates a tailoring of the representation languages for the problem at hand which is an advantage over traditional transparent techniques as decision trees that are designed to optimize a specific representation that cannot be changed without changing the algorithm. Experiments in study 8.5 show that the G-REX LINREG representation, that is more suited for the domain, clearly outperforms the other representations. This is an example how tailoring the representation language can help to increase the performance. It should also be noted that tailoring of the representations language could also be used to increase the comprehensibility of the extracted rules by creating a representation that are more to the decisions makers preferences. When compared to MREG that has a similar representation which has a similar representation, LINREG is also significantly better in regards of GMRAE. The same results are also achieved in study 8.6 where LINREG is significantly better than MREG in regards of GMRAE and also outperforms Tree Fit and G-REX NORM. However when evaluated on MUPAE the experiments have no clear winner; Tree Fit is still significantly better than MREG and the G-REX NORM representation is better than Tree Fit but none of the techniques stands out. On the other hand, each of the tested G-REX representations is clearly better than the correspondent 133
134 transparent model. In any case G-REX is well suited to extract rule for estimation tasks and the possibility to tailor the representation is yet another advantage that could be used for the benefit of a decision maker. 4. Analyzing the effects of optimizing alternative performance metrics. In the same way that most techniques are designed for a single representation also designed to optimize a single performance metric, most often accuracy. However researcher has pointed out that accuracy has several deficiencies. There are also cases when accuracy is not the performance that is sought but an ability to rank or estimate a probability. Yet again G-REXs GP foundation makes it possible to optimize arbitrary performance metric or even combination of metrics. Study 8.2 not surprisingly shows that it is better to optimize the metric that is supposed to be optimized than something else, i.e. if AUC is sought AUC should be optimized. This might seem like a very obvious result but this should be seen in the light of that most techniques only optimize a single predefined metric and thus the choice of technique decides the optimization criterion. An example of the advantage of this feature is that new performance metrics easily can be created and optimized. Study 8.2 introduces a new performance metric called Brevity, which is supposed to capture how much effort a model requires from a decision maker to make a prediction. The idea is that simple general examples should be classified based a very small set of condition while the more complex expressions are reserved for more uncharacteristic cases. In the experiments where G-REX optimizes brevity it produces rules with comparable accuracy but significantly better brevity. The produced rules classifies the majority of the instances with one or two conditions an only a few instance are classified with more complex expressions. In this way a performance metric can be used to govern other properties than accuracy that can be more important for the current problem. 5. Evaluate techniques for providing case based uncertainty estimations from ensembles. Ensembles as all models rarely show uniform distribution of its error since some instances are easier to predict than others. If the difficulty could be estimated for each instance, this information could be used to minimize the need for rule extraction by pin pointing the instances that would need special attention from the decision maker. Problematic instances could be presented to 134
135 the decision maker with an attached explanation based on the extracted model and everything else could be handled automatically. Study 8.3 shows that probability estimation can be done with great success in several ways. When the best technique which were based on GA, ranked the instance according to its confidence for ensembles predictions; the top 50% (the easy instance) had a mean error that was only 62% of the mean error for all instances. It is apparently quite possible to estimates the ensembles confidence in each prediction and hence this approach of pin pointing troublesome instance is feasible. However a comprehensible model is always important since another reason for adjusting a prediction is that something outside the model changed and hence is not accounted for in the prediction. 9.1 Concluding remarks Based on the conclusions above rule extraction from ensembles should definitely be considered when building models for domains with poor data quality. Ensemble techniques ensures robust high performing models and can also be used to indicates which instances that should be analyzed further by domain experts. Here rule extraction can provide the domain experts with different possible explanations (G-REX is non deterministic) which can be used as a basis for further analysis. This method minimizes the attention of a decision maker while ensuring a safer procedure for harder predictions. 135
136
137 10 Future Work Information Fusion A general problem for all ML-techniques is how rare and new events should be handled. As ML requires several examples to find a pattern it is impossible to do this automatically and instead domain experts have to make manual adjustments. A comprehensible model is crucial as a rational base for such adjustments but previous research have also shown that the final adjustments between domain experts and the ML-model are best done mechanically. However it is not clear how the fusion of the ML-model and the Expert should be done in the best way and this is therefore an interesting task for future research. A domain experts could either be allowed to adjust a prediction or make an own prediction that could be fused with the original prediction. Another option is to let the domain expert continuously judge the importance of event and use these estimations as an input when creating the model. It is also unclear how new trends should be incorporated if there is not enough evidence of the trend in the data. Parameter evaluation Another area for use of predictive models is parameter evaluation. One example is marketing mix models where the ML-model is used to estimate the effect of different marketing efforts. Here it is also crucial that domain knowledge can be incorporated if maximal performance is to be achieved. An interesting difference is that in this case the domain expert can be used to give retroactive explanations of events that were not captured by the model. These explanations could then be used together with the original data to create a new model with higher performance thus fusing the expert knowledge with the data. Utilize inconsistency As seen in study 8.4 inconsistency of a rule inducer could actually be used to create improved probability estimates. It would be interesting to continue this work and evaluate if the inconsistency could be used in other ways. One example could be to create a metric that would give a decision maker a way to realize the level of ambiguity present in a dataset. Thus it would be possible to better judge which confidence that should be put in models created from the dataset. It could also be possible relate the ambiguity to a certain part of the found relationship. This could possibly lead to new insight of which data that needs to be added to decrease the ambiguity and improve the performance of the extracted model. 137
138 Another way to utilize the inconsistency is to simply create ensembles of accurate inconsistent models. Here the inconsistency would be a way of creating diversity without modifying the data with weights as in boosting or creating bootstraps as in baging. 138
139 11 References. Richard Roiger and Michael Geatz. Data mining : a tutorial-based primer, Addison Wesley, Boston, AA99. Thomas L. Ainscough and Jay E. Aronson. An empirical investigation and comparison of neural networks and regression for scanner data analysis. Journal of Retailing and Consumer Services, 6, , AC92. J. Scott Armstrong and Fred Collopy. Error Measures For Generalizing About Forecasting Methods: Empirical Comparisons. International Journal of Forecasting,, 8, 69-80, ADT95. Robert Andrews, Joachim Diederich and Alan B. Tickle. Survey and critique of techniques for extracting rules from trained artificial neural networks. Knowledge-Based Systems, 8, , AG95. R. Andrews and S. Geva, "RULEX and CEBP networks as a basis for a rule refinement system," presented at AISB 95, tenth biennial conference on AI and Cognitive Science, Sheffield, England, ANO05. Sten F. Andler, Lars Niklasson, Bjorn Olsson, Anne Persson, Leo J. de Vin, Benkt Wangler, Tom Ziemke and Tomas Planstedt, "Information Fusion from Databases, Sensors and Simulations: A Collaborative Research Program," ARM01. J. Scott Armstrong. Principles of forecasting : a handbook for researchers and practitioners, Kluwer Academic, Boston, Mass. ; London, BFO+83. L. Breiman, J. H. Friedman, R. A. Olshen and C. J. Stone. Classification and Regression Trees, Wadsworth, BL04. Michael J. A. Berry and Gordon Linoff. Data mining techniques : for marketing, sales and customer relationship management, Wiley, New York ; Chichester, BL88. D. S. Broomhead and D. Lowe. Multivariable functional interpolation and adaptive networks. Complex Systems, 2, , BM98. C. L. Blake and C. J. Merz University of California, BPR02. R. Breton, S. Paradis and J. Roy, "Command Decision Support Interface (CODSI) for human factors and display concept validation," presented at Fifth International Conference on Information Fusion Bre96. Leo Breiman. Bagging Predictors. Machine Learning, 24, , Brie50. Glenn W. Brier. Verification of Forecasts Expressed in Terms of Probability. Monthly Weahter Review, 78, CCK+. P. Chapman, C. Clinton, R. Kerber, T. Khabaza, T. Reinartz, C. Shearer and Wirth R., CGCP05. Subhash Chalia, Tauseef Gulrez, Zenon Chaczko and T.N. Paranesha, "Opportunistic Information Fusion: A New Paradigm for Next Generation Networked Sensing Systems," presented at 8th International Conference on Information Fusion, CS96. Mark W. Craven and Jude W. Shavlik. Extracting Tree-Structured Representations of Trained Networks. Advances in Neural Information Processing Systems, 8, 24-30, CS99. Mark Craven and Jude Shavlik University of Wisonsin Machine Learning Research Group, Das01. B. V. Dasarathy. Information Fusion - what, where, why, when, and how? Information Fusion, 2, 75-76, DBS77. Randall Davis, Bruce Buchanan and Edward Shortliffe. Production rules as a representation for a knowledge-based consultation program. Artificial Intelligence, 8, 15-45,
140 Dem06. J. Demsar. Statistical Comparisons of Classifiers over Multiple Data Sets. Journal of Machine Learning Research, 7, 1-30, Dun03. Margaret H. Dunham. Data mining : introductory and advanced topics, Prentice Hall/Pearson Education, Upper Saddle River, NJ, Dun61. J. Dunn. Multiple comparisons among means. Journal of American Statistical Association, 56, 52-64, Ehr97. Martin Ehrendorfer. Predicting the uncertainty of numerical weather forecasts: a review. Meteorol. Zeitschrift, N.. 6, , Faw06. T. Fawcett. An introduction to ROC analysis. Pattern Recognition Letters, 27, , Fri37. M. Friedman. The Use of Ranks to Avoid the Assumption of Normality. Journal of American Statistical Association, 32, , GB05. Kai Goebel and Piero Bonissone, "Prognostic Information Fusion for Constant Load Systems," presented at Information Fusion, GKP99. Thomas S. Gruca, Bruce R. Klemz and E. Ann Furr Petersen. Mining Sales Data using a Neural Network Model of Market Response. SIGKDD Explorations, 1, 39-43, Goo02. Paul Goodwin. Integrating management judgment and statistical methods to improve short-term forecasts. Omega, 30, , Gra05. Tim Grant In Proceedings of the 2005 annual research conference of the South African institute of computer scientists and information technologists on IT research in developing countriessouth African Institute for Computer Scientists and Information Technologists, White River, South Africa, HBD96. Martin T. Hagan, Mark H. Beale and Howard B. Demuth. Neural network design, PWS Pub., Boston ; London, HBV06. J. Huysmans, B. Baesens and J. Vanthienen (Ed, Leuven, K. U.) FETEW Research Report KBI_0612, HL01. David L. Hall and James Llinas. Handbook of multisensor data fusion / [Elektronisk resurs], CRC Press, Boca Raton, Fla., Hol75. J. H. Holland. Adaptation in Natural and Artificial Systems, University of Michigan Press, Hor91. Kurt Hornik. Approximation capabilities of multilayer feedforward networks. Neural Networks, 4, , Hrus93. Harald Hruschka. Determining market response functions by neural network modeling: A comparison to econometric techniques. European Journal of Operational Research, 66, 27-35, HSY94. S. Hashem, B. Schmeiser and Y. Yih, "Optimal linear combinations of neural networks: an overview," presented at IEEE International Conference on Neural Networks, HW01. Se June Hong and Sholom M. Weiss. Advances in predictive models for data mining. Pattern Recognition Letters, 22, 55-61, JKN03. Ulf Johansson, Rikard Konig and Lars Niklasson In Joint 13th International Conference on Artificial Neural Networks and 10th International Conference on Neural Information Processing, ICANN/ICONIP 2003Istanbul, Turkey, JKN04. Ulf Johansson, Rikard Konig and Lars Niklasson, "The truth is in there - Rule extraction from opaque models using genetic programming," presented at FLAIRS Proceedings of the Seventeenth International Florida Artificial Intelligence Research Society Conference, JKN05. U. Johansson, R. Konig and L. Niklasson, "Automatically balancing accuracy and comprehensibility in predictive modeling," presented at 8th International Conference on Information Fusion (FUSION), pp. JKN07. Ulf Johansson, Rikard Konig and Lars Niklasson, "Inconsistency Friend or Foe," presented at IJCNN, International Joint Conference on Neural Networks,
141 JLK+06. Ulf Johansson, Tuve Lofstrom, Rikard Konig, Cecilia Sonstrod and Lars Niklasson, "Rule Extraction from Opaque Models - A Slightly Different Perspective," presented at ICMLA, JNK04. Ulf Johansson, Lars Niklasson and Rikard Konig, "Accuracy vs. comprehensibility in data mining models," presented at Seventh International Conference on Information Fusion, Stockholm, Sweden, Joh07. Ulf Johansson. Obtaining accurate and comprehensible data mining models : an evolutionary approach, Department of Computer and Information Science, Linköpings universitet, Linköping, JSN04. Ulf Johansson, Cecilia Sonstrod and Lars Niklasson, "Why Rule Extraction Matters," presented at SEA, Software Engineering and Applications, Kas80. G. V. Kass. An Exploratory Technique for Investigating Large Quantities of Categorical Data. Applied Statistics, 29, , KH90. V. Kumar and Timothy B. Heath. A comparative study of market share models using disaggregate data. International Journal of Forecasting, 6, , Kim06. Yong Seog Kim. Toward a successful CRM: variable selection, sampling, and ensemble. Decision Support Systems, 41, , KJN06. Rikard Konig, Ulf Johansson and Lars Niklasson. Increasing rule extraction comprehensibility. International Journal of Information Technology and Intelligent Computing, 1, , KJN07. Rikard Konig, Ulf Johansson and Lars Niklasson, "Genetic Programming a Tool for Flexible Rule Extraction," presented at IEEE Congress on Evolutionary Computation, CEC, KJN07. Rikard Konig, Ulf Johansson and Lars Niklasson, "Instance ranking using ensemble spread," presented at DMIN, The 2007 international conference on data mining, Las Vegas Nevada, USA, KJN08. R. Konig, U. Johansson and L. Niklasson, "Using Genetic Programming to Increase Rule Quality," presented at Proceedings of the Twenty-First International FLAIRS Conference, Coconut Grove, Florida, USA, Koz92. John R. Koza. Genetic programming : on the programming of computers by means of natural selection, MIT, Cambridge, Mass., KV95. Anders Krogh and Jesper Vedelsby. Neural Network Ensembles, Cross Validation, and Active Learning. Advances in Neural Information Processing Systems, 7, , LCV02. C.A.M. Lima, A.L.V. Coelho and F.J. Von Zuben, "Ensembles of support vector machines for regression problems," presented at International Joint Conference on Neural Networks, LZ01. X. Rong Li and Zhanlue Zhao, "Measures of Performance for evaluation of estimators and Filters," presented at Proc. SPIE Conference on Signal and Data Processing, LZ05. X. Rong Li and Zhanlue Zhao, "Relative Error Measures for Evaluation of Estimation Algorithms," presented at 7th International Conference on Information Fusion (FUSION), Mak93. Spyros Makridakis. Accuracy measures: theoretical and practical concerns. International Journal of Forecasting, 9, , MC05. B. L. Milenova and M. M. Campos, "Mining high-dimensional data for information fusion: a database-centric approach," presented at Information Fusion, th International Conference on, pp. MD02. D. Margineantu and T. Dietterich. Improved Class Probability Estimates from Decision Tree Models. Nonlinear Estimation and Classification; Lecture Notes in Statistics, 171, , MF04. Saar-Tsechansky Maytal and Provost Foster. Active Sampling for Class Probability Estimation and Ranking. Mach. Learn., 54, , Mit97. Tom Michael Mitchell. Machine learning, McGraw-Hill, New York, Nem63. P. B. Nemenyi Princeton University,
142 PD03. Foster Provost and Pedro Domingos. Tree Induction for Probability-Based Ranking. Mach. Learn., 52, , PK01. A Papagelis and D. Kalles In 18th International Conference on Machine LearningWilliamstown, MA, pp , PTR98. F. Provost, T. Fawcett and R. Kohavi In 15th International Conf. on Machine LearningMorgan Kaufmann, San Franciscio, CA, pp , PTR98. Foster Provost, T. Fawcett and R. Kohavi In 15th International Conf. on Machine LearningMorgan Kaufmann, San Franciscio, CA, pp , Qui86. J. R. Quinlan. Induction of Decision Trees. Mach. Learn., 1, , Qui93. J. Ross Quinlan. C4.5: programs for machine learning, Morgan Kaufmann Publishers Inc., RG03. Richard Roiger and Michael Geatz. Data mining: a tutorial-based primer, Addison Wesley, Boston, Ris78. J Rissanen. Modeling by shortest data description. Automatica, 14, RM86. David E. Rumelhart and James L. McClelland. Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations, MIT Press, RS03. M. S. Roulston and L. A. Smith. Combining dynamical and statistical ensembles Tellus A, 55, 16-30, SBW99. Alan N. Steinberg, Christopher L. Bowman and Franklin E. White, "Revisions to the JDL data fusion model," presented at SPIE, Sch89. R. E. Schapire, "The strength of weak learnability," Research Triangle Park, NC, USA, Sch94. C. Schaffer, "A conservation law for generalization performance," presented at Eleventh International Conference on Machine Learning, New Brunswick, SHBB03. John Salemo, Mike Hinman, Doug Boulware and Paul Bello, "Information Fusion for Situational Awareness," presented at Sixth International Conference of Information Fusion, Shes00. J. Sheskin. Handbook of parametric and nonparametric statistical procedures., Chapman & Hall/CRC, SK02. L Stefanova and T. N. Krishnamurti. Interpretation of Seasonal Climate Forecast Using Brier Skill Score, The Florida State University Superensemble, and the AMIP-I Dataset. American Meteorological Society, 15, , Spe90. Donald F. Specht. Probabilistic neural networks. Neural Networks, 3, , TCM01. Zoltan Toth, Yuejian Zhu and Timothy Marchok. The Use of Ensembles to Identify Forecasts with Small and Large Uncertainty. Weather and Forecasting, 16, , Thr93. Sebastian B. Thrun, TS93. G. Towell and J. Shavlik. The Extraction of Refined Rules from Knowledge-Based Neural Networks Machine Learning, 13, , TSK06. Pang-Ning Tan, Michael Steinbach and Vipin Kumar. Introduction to data mining, Pearson Addison Wesley, Boston, UCK+03. Ulf Johansson, Cecilia Sonstrod, R. Konig and Lars Niklasson, "Neural networks and rule extraction for prediction and explanation in the marketing domain," presented at Proceedings of the International Joint Conference on Neural Networks, WF05. Ian H. Witten and Eibe Frank. Data mining : practical machine learning tools and techniques, Morgan Kaufmann, Amsterdam, Wil45. F. Wilcoxon. Individual comparisons by ranking methods. Biometrics, 1, 80-83, Zha04. G. P. Zhang. Neural networks in business forecasting, Idea Group Publishing, Hershey, PA, Zhu04. Yuejian Zhu. Probabilistic forecasts and evaluations based on global ensemble prediction system. World Scientific Sreies on Meteorlogy of east Asia, 3, , Zhu05. Yuejian Zhu. Ensemble Forecast: A New Approach to Uncertainty and Predictability. Advances in Atmospheric Sciences, 22, ,
143 ZPH98. Guoqiang Zhang, Eddy B. Patuwo and Michael Y. Hu. Forecasting with artificial neural networks:: The state of the art. International Journal of Forecasting, 14, 35-62,
Predictive Techniques and Methods for Decision Support in Situations with Poor Data Quality
Licentiate Thesis Predictive Techniques and Methods for Decision Support in Situations with Poor Data Quality RIKARD KÖNIG Technology UNIVERSITY OF BORÅS School of Business and Informatics UNIVERSITY OF
Data Mining - Evaluation of Classifiers
Data Mining - Evaluation of Classifiers Lecturer: JERZY STEFANOWSKI Institute of Computing Sciences Poznan University of Technology Poznan, Poland Lecture 4 SE Master Course 2008/2009 revised for 2010
Database Marketing, Business Intelligence and Knowledge Discovery
Database Marketing, Business Intelligence and Knowledge Discovery Note: Using material from Tan / Steinbach / Kumar (2005) Introduction to Data Mining,, Addison Wesley; and Cios / Pedrycz / Swiniarski
Gerard Mc Nulty Systems Optimisation Ltd [email protected]/0876697867 BA.,B.A.I.,C.Eng.,F.I.E.I
Gerard Mc Nulty Systems Optimisation Ltd [email protected]/0876697867 BA.,B.A.I.,C.Eng.,F.I.E.I Data is Important because it: Helps in Corporate Aims Basis of Business Decisions Engineering Decisions Energy
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
not possible or was possible at a high cost for collecting the data.
Data Mining and Knowledge Discovery Generating knowledge from data Knowledge Discovery Data Mining White Paper Organizations collect a vast amount of data in the process of carrying out their day-to-day
Chapter 12 Discovering New Knowledge Data Mining
Chapter 12 Discovering New Knowledge Data Mining Becerra-Fernandez, et al. -- Knowledge Management 1/e -- 2004 Prentice Hall Additional material 2007 Dekai Wu Chapter Objectives Introduce the student to
Data Mining Solutions for the Business Environment
Database Systems Journal vol. IV, no. 4/2013 21 Data Mining Solutions for the Business Environment Ruxandra PETRE University of Economic Studies, Bucharest, Romania [email protected] Over
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Information Visualization WS 2013/14 11 Visual Analytics
1 11.1 Definitions and Motivation Lot of research and papers in this emerging field: Visual Analytics: Scope and Challenges of Keim et al. Illuminating the path of Thomas and Cook 2 11.1 Definitions and
Basic Concepts in Research and Data Analysis
Basic Concepts in Research and Data Analysis Introduction: A Common Language for Researchers...2 Steps to Follow When Conducting Research...3 The Research Question... 3 The Hypothesis... 4 Defining the
Why Taking This Course? Course Introduction, Descriptive Statistics and Data Visualization. Learning Goals. GENOME 560, Spring 2012
Why Taking This Course? Course Introduction, Descriptive Statistics and Data Visualization GENOME 560, Spring 2012 Data are interesting because they help us understand the world Genomics: Massive Amounts
Chapter 6. The stacking ensemble approach
82 This chapter proposes the stacking ensemble approach for combining different data mining classifiers to get better performance. Other combination techniques like voting, bagging etc are also described
Data Mining Algorithms Part 1. Dejan Sarka
Data Mining Algorithms Part 1 Dejan Sarka Join the conversation on Twitter: @DevWeek #DW2015 Instructor Bio Dejan Sarka ([email protected]) 30 years of experience SQL Server MVP, MCT, 13 books 7+ courses
DATA MINING TECHNIQUES AND APPLICATIONS
DATA MINING TECHNIQUES AND APPLICATIONS Mrs. Bharati M. Ramageri, Lecturer Modern Institute of Information Technology and Research, Department of Computer Application, Yamunanagar, Nigdi Pune, Maharashtra,
Data quality in Accounting Information Systems
Data quality in Accounting Information Systems Comparing Several Data Mining Techniques Erjon Zoto Department of Statistics and Applied Informatics Faculty of Economy, University of Tirana Tirana, Albania
An Overview of Knowledge Discovery Database and Data mining Techniques
An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,
Knowledge Discovery from patents using KMX Text Analytics
Knowledge Discovery from patents using KMX Text Analytics Dr. Anton Heijs [email protected] Treparel Abstract In this white paper we discuss how the KMX technology of Treparel can help searchers
International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014
RESEARCH ARTICLE OPEN ACCESS A Survey of Data Mining: Concepts with Applications and its Future Scope Dr. Zubair Khan 1, Ashish Kumar 2, Sunny Kumar 3 M.Tech Research Scholar 2. Department of Computer
Comparison of K-means and Backpropagation Data Mining Algorithms
Comparison of K-means and Backpropagation Data Mining Algorithms Nitu Mathuriya, Dr. Ashish Bansal Abstract Data mining has got more and more mature as a field of basic research in computer science and
DMDSS: Data Mining Based Decision Support System to Integrate Data Mining and Decision Support
DMDSS: Data Mining Based Decision Support System to Integrate Data Mining and Decision Support Rok Rupnik, Matjaž Kukar, Marko Bajec, Marjan Krisper University of Ljubljana, Faculty of Computer and Information
SPATIAL DATA CLASSIFICATION AND DATA MINING
, pp.-40-44. Available online at http://www. bioinfo. in/contents. php?id=42 SPATIAL DATA CLASSIFICATION AND DATA MINING RATHI J.B. * AND PATIL A.D. Department of Computer Science & Engineering, Jawaharlal
Cross Validation. Dr. Thomas Jensen Expedia.com
Cross Validation Dr. Thomas Jensen Expedia.com About Me PhD from ETH Used to be a statistician at Link, now Senior Business Analyst at Expedia Manage a database with 720,000 Hotels that are not on contract
CRISP - DM. Data Mining Process. Process Standardization. Why Should There be a Standard Process? Cross-Industry Standard Process for Data Mining
Mining Process CRISP - DM Cross-Industry Standard Process for Mining (CRISP-DM) European Community funded effort to develop framework for data mining tasks Goals: Cross-Industry Standard Process for Mining
Data Mining and Knowledge Discovery in Databases (KDD) State of the Art. Prof. Dr. T. Nouri Computer Science Department FHNW Switzerland
Data Mining and Knowledge Discovery in Databases (KDD) State of the Art Prof. Dr. T. Nouri Computer Science Department FHNW Switzerland 1 Conference overview 1. Overview of KDD and data mining 2. Data
Data Mining. Nonlinear Classification
Data Mining Unit # 6 Sajjad Haider Fall 2014 1 Nonlinear Classification Classes may not be separable by a linear boundary Suppose we randomly generate a data set as follows: X has range between 0 to 15
Machine Learning and Data Mining. Fundamentals, robotics, recognition
Machine Learning and Data Mining Fundamentals, robotics, recognition Machine Learning, Data Mining, Knowledge Discovery in Data Bases Their mutual relations Data Mining, Knowledge Discovery in Databases,
Knowledge Discovery and Data Mining
Knowledge Discovery and Data Mining Unit # 6 Sajjad Haider Fall 2014 1 Evaluating the Accuracy of a Classifier Holdout, random subsampling, crossvalidation, and the bootstrap are common techniques for
Using reporting and data mining techniques to improve knowledge of subscribers; applications to customer profiling and fraud management
Using reporting and data mining techniques to improve knowledge of subscribers; applications to customer profiling and fraud management Paper Jean-Louis Amat Abstract One of the main issues of operators
PREDICTING STOCK PRICES USING DATA MINING TECHNIQUES
The International Arab Conference on Information Technology (ACIT 2013) PREDICTING STOCK PRICES USING DATA MINING TECHNIQUES 1 QASEM A. AL-RADAIDEH, 2 ADEL ABU ASSAF 3 EMAN ALNAGI 1 Department of Computer
Chapter 7. One-way ANOVA
Chapter 7 One-way ANOVA One-way ANOVA examines equality of population means for a quantitative outcome and a single categorical explanatory variable with any number of levels. The t-test of Chapter 6 looks
A Review of Data Mining Techniques
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 4, April 2014,
Data Mining Applications in Higher Education
Executive report Data Mining Applications in Higher Education Jing Luan, PhD Chief Planning and Research Officer, Cabrillo College Founder, Knowledge Discovery Laboratories Table of contents Introduction..............................................................2
Cross-Validation. Synonyms Rotation estimation
Comp. by: BVijayalakshmiGalleys0000875816 Date:6/11/08 Time:19:52:53 Stage:First Proof C PAYAM REFAEILZADEH, LEI TANG, HUAN LIU Arizona State University Synonyms Rotation estimation Definition is a statistical
How To Use Neural Networks In Data Mining
International Journal of Electronics and Computer Science Engineering 1449 Available Online at www.ijecse.org ISSN- 2277-1956 Neural Networks in Data Mining Priyanka Gaur Department of Information and
Benchmarking of different classes of models used for credit scoring
Benchmarking of different classes of models used for credit scoring We use this competition as an opportunity to compare the performance of different classes of predictive models. In particular we want
Data, Measurements, Features
Data, Measurements, Features Middle East Technical University Dep. of Computer Engineering 2009 compiled by V. Atalay What do you think of when someone says Data? We might abstract the idea that data are
In this presentation, you will be introduced to data mining and the relationship with meaningful use.
In this presentation, you will be introduced to data mining and the relationship with meaningful use. Data mining refers to the art and science of intelligent data analysis. It is the application of machine
DATA MINING TECHNIQUES SUPPORT TO KNOWLEGDE OF BUSINESS INTELLIGENT SYSTEM
INTERNATIONAL JOURNAL OF RESEARCH IN COMPUTER APPLICATIONS AND ROBOTICS ISSN 2320-7345 DATA MINING TECHNIQUES SUPPORT TO KNOWLEGDE OF BUSINESS INTELLIGENT SYSTEM M. Mayilvaganan 1, S. Aparna 2 1 Associate
The Science and Art of Market Segmentation Using PROC FASTCLUS Mark E. Thompson, Forefront Economics Inc, Beaverton, Oregon
The Science and Art of Market Segmentation Using PROC FASTCLUS Mark E. Thompson, Forefront Economics Inc, Beaverton, Oregon ABSTRACT Effective business development strategies often begin with market segmentation,
Bootstrapping Big Data
Bootstrapping Big Data Ariel Kleiner Ameet Talwalkar Purnamrita Sarkar Michael I. Jordan Computer Science Division University of California, Berkeley {akleiner, ameet, psarkar, jordan}@eecs.berkeley.edu
Making Sense of the Mayhem: Machine Learning and March Madness
Making Sense of the Mayhem: Machine Learning and March Madness Alex Tran and Adam Ginzberg Stanford University [email protected] [email protected] I. Introduction III. Model The goal of our research
Evaluation & Validation: Credibility: Evaluating what has been learned
Evaluation & Validation: Credibility: Evaluating what has been learned How predictive is a learned model? How can we evaluate a model Test the model Statistical tests Considerations in evaluating a Model
Big Data: Rethinking Text Visualization
Big Data: Rethinking Text Visualization Dr. Anton Heijs [email protected] Treparel April 8, 2013 Abstract In this white paper we discuss text visualization approaches and how these are important
AP Physics 1 and 2 Lab Investigations
AP Physics 1 and 2 Lab Investigations Student Guide to Data Analysis New York, NY. College Board, Advanced Placement, Advanced Placement Program, AP, AP Central, and the acorn logo are registered trademarks
Gerry Hobbs, Department of Statistics, West Virginia University
Decision Trees as a Predictive Modeling Method Gerry Hobbs, Department of Statistics, West Virginia University Abstract Predictive modeling has become an important area of interest in tasks such as credit
Lecture 2: Descriptive Statistics and Exploratory Data Analysis
Lecture 2: Descriptive Statistics and Exploratory Data Analysis Further Thoughts on Experimental Design 16 Individuals (8 each from two populations) with replicates Pop 1 Pop 2 Randomly sample 4 individuals
Knowledge Discovery and Data Mining
Knowledge Discovery and Data Mining Unit # 11 Sajjad Haider Fall 2013 1 Supervised Learning Process Data Collection/Preparation Data Cleaning Discretization Supervised/Unuspervised Identification of right
THE HYBRID CART-LOGIT MODEL IN CLASSIFICATION AND DATA MINING. Dan Steinberg and N. Scott Cardell
THE HYBID CAT-LOGIT MODEL IN CLASSIFICATION AND DATA MINING Introduction Dan Steinberg and N. Scott Cardell Most data-mining projects involve classification problems assigning objects to classes whether
Measurement Information Model
mcgarry02.qxd 9/7/01 1:27 PM Page 13 2 Information Model This chapter describes one of the fundamental measurement concepts of Practical Software, the Information Model. The Information Model provides
Knowledge Discovery and Data Mining
Knowledge Discovery and Data Mining Unit # 10 Sajjad Haider Fall 2012 1 Supervised Learning Process Data Collection/Preparation Data Cleaning Discretization Supervised/Unuspervised Identification of right
A STUDY OF DATA MINING ACTIVITIES FOR MARKET RESEARCH
205 A STUDY OF DATA MINING ACTIVITIES FOR MARKET RESEARCH ABSTRACT MR. HEMANT KUMAR*; DR. SARMISTHA SARMA** *Assistant Professor, Department of Information Technology (IT), Institute of Innovation in Technology
NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )
Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates
Knowledge Discovery and Data Mining. Structured vs. Non-Structured Data
Knowledge Discovery and Data Mining Unit # 2 1 Structured vs. Non-Structured Data Most business databases contain structured data consisting of well-defined fields with numeric or alphanumeric values.
The Predictive Data Mining Revolution in Scorecards:
January 13, 2013 StatSoft White Paper The Predictive Data Mining Revolution in Scorecards: Accurate Risk Scoring via Ensemble Models Summary Predictive modeling methods, based on machine learning algorithms
APPENDIX N. Data Validation Using Data Descriptors
APPENDIX N Data Validation Using Data Descriptors Data validation is often defined by six data descriptors: 1) reports to decision maker 2) documentation 3) data sources 4) analytical method and detection
Comparing the Results of Support Vector Machines with Traditional Data Mining Algorithms
Comparing the Results of Support Vector Machines with Traditional Data Mining Algorithms Scott Pion and Lutz Hamel Abstract This paper presents the results of a series of analyses performed on direct mail
Multiple Linear Regression in Data Mining
Multiple Linear Regression in Data Mining Contents 2.1. A Review of Multiple Linear Regression 2.2. Illustration of the Regression Process 2.3. Subset Selection in Linear Regression 1 2 Chap. 2 Multiple
Data Mining Framework for Direct Marketing: A Case Study of Bank Marketing
www.ijcsi.org 198 Data Mining Framework for Direct Marketing: A Case Study of Bank Marketing Lilian Sing oei 1 and Jiayang Wang 2 1 School of Information Science and Engineering, Central South University
Recall this chart that showed how most of our course would be organized:
Chapter 4 One-Way ANOVA Recall this chart that showed how most of our course would be organized: Explanatory Variable(s) Response Variable Methods Categorical Categorical Contingency Tables Categorical
Overview. Evaluation Connectionist and Statistical Language Processing. Test and Validation Set. Training and Test Set
Overview Evaluation Connectionist and Statistical Language Processing Frank Keller [email protected] Computerlinguistik Universität des Saarlandes training set, validation set, test set holdout, stratification
Data Mining and Neural Networks in Stata
Data Mining and Neural Networks in Stata 2 nd Italian Stata Users Group Meeting Milano, 10 October 2005 Mario Lucchini e Maurizo Pisati Università di Milano-Bicocca [email protected] [email protected]
Machine Learning Capacity and Performance Analysis and R
Machine Learning and R May 3, 11 30 25 15 10 5 25 15 10 5 30 25 15 10 5 0 2 4 6 8 101214161822 0 2 4 6 8 101214161822 0 2 4 6 8 101214161822 100 80 60 40 100 80 60 40 100 80 60 40 30 25 15 10 5 25 15 10
Simple Predictive Analytics Curtis Seare
Using Excel to Solve Business Problems: Simple Predictive Analytics Curtis Seare Copyright: Vault Analytics July 2010 Contents Section I: Background Information Why use Predictive Analytics? How to use
Use of Data Mining in the field of Library and Information Science : An Overview
512 Use of Data Mining in the field of Library and Information Science : An Overview Roopesh K Dwivedi R P Bajpai Abstract Data Mining refers to the extraction or Mining knowledge from large amount of
Sanjeev Kumar. contribute
RESEARCH ISSUES IN DATAA MINING Sanjeev Kumar I.A.S.R.I., Library Avenue, Pusa, New Delhi-110012 [email protected] 1. Introduction The field of data mining and knowledgee discovery is emerging as a
Improving the Performance of Data Mining Models with Data Preparation Using SAS Enterprise Miner Ricardo Galante, SAS Institute Brasil, São Paulo, SP
Improving the Performance of Data Mining Models with Data Preparation Using SAS Enterprise Miner Ricardo Galante, SAS Institute Brasil, São Paulo, SP ABSTRACT In data mining modelling, data preparation
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING USING NEURAL NETWORKS JothiKumar.R 1, Sivabalan.R.V 2 1 Research scholar, Noorul Islam University, Nagercoil, India Assistant Professor, Adhiparasakthi College
Machine Learning in Hospital Billing Management. 1. George Mason University 2. INOVA Health System
Machine Learning in Hospital Billing Management Janusz Wojtusiak 1, Che Ngufor 1, John M. Shiver 1, Ronald Ewald 2 1. George Mason University 2. INOVA Health System Introduction The purpose of the described
IBM SPSS Statistics 20 Part 4: Chi-Square and ANOVA
CALIFORNIA STATE UNIVERSITY, LOS ANGELES INFORMATION TECHNOLOGY SERVICES IBM SPSS Statistics 20 Part 4: Chi-Square and ANOVA Summer 2013, Version 2.0 Table of Contents Introduction...2 Downloading the
Dimensionality Reduction: Principal Components Analysis
Dimensionality Reduction: Principal Components Analysis In data mining one often encounters situations where there are a large number of variables in the database. In such situations it is very likely
Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm
Mgt 540 Research Methods Data Analysis 1 Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm http://web.utk.edu/~dap/random/order/start.htm
Learning is a very general term denoting the way in which agents:
What is learning? Learning is a very general term denoting the way in which agents: Acquire and organize knowledge (by building, modifying and organizing internal representations of some external reality);
STATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
On the effect of data set size on bias and variance in classification learning
On the effect of data set size on bias and variance in classification learning Abstract Damien Brain Geoffrey I Webb School of Computing and Mathematics Deakin University Geelong Vic 3217 With the advent
Perspectives on Data Mining
Perspectives on Data Mining Niall Adams Department of Mathematics, Imperial College London [email protected] April 2009 Objectives Give an introductory overview of data mining (DM) (or Knowledge Discovery
Clustering & Visualization
Chapter 5 Clustering & Visualization Clustering in high-dimensional databases is an important problem and there are a number of different clustering paradigms which are applicable to high-dimensional data.
Statistics 215b 11/20/03 D.R. Brillinger. A field in search of a definition a vague concept
Statistics 215b 11/20/03 D.R. Brillinger Data mining A field in search of a definition a vague concept D. Hand, H. Mannila and P. Smyth (2001). Principles of Data Mining. MIT Press, Cambridge. Some definitions/descriptions
Feature vs. Classifier Fusion for Predictive Data Mining a Case Study in Pesticide Classification
Feature vs. Classifier Fusion for Predictive Data Mining a Case Study in Pesticide Classification Henrik Boström School of Humanities and Informatics University of Skövde P.O. Box 408, SE-541 28 Skövde
Divide-n-Discover Discretization based Data Exploration Framework for Healthcare Analytics
for Healthcare Analytics Si-Chi Chin,KiyanaZolfaghar,SenjutiBasuRoy,AnkurTeredesai,andPaulAmoroso Institute of Technology, The University of Washington -Tacoma,900CommerceStreet,Tacoma,WA980-00,U.S.A.
Data Isn't Everything
June 17, 2015 Innovate Forward Data Isn't Everything The Challenges of Big Data, Advanced Analytics, and Advance Computation Devices for Transportation Agencies. Using Data to Support Mission, Administration,
Three types of messages: A, B, C. Assume A is the oldest type, and C is the most recent type.
Chronological Sampling for Email Filtering Ching-Lung Fu 2, Daniel Silver 1, and James Blustein 2 1 Acadia University, Wolfville, Nova Scotia, Canada 2 Dalhousie University, Halifax, Nova Scotia, Canada
White Paper. Data Mining for Business
White Paper Data Mining for Business January 2010 Contents 1. INTRODUCTION... 3 2. WHY IS DATA MINING IMPORTANT?... 3 FUNDAMENTALS... 3 Example 1...3 Example 2...3 3. OPERATIONAL CONSIDERATIONS... 4 ORGANISATIONAL
DATA MINING TECHNOLOGY. Keywords: data mining, data warehouse, knowledge discovery, OLAP, OLAM.
DATA MINING TECHNOLOGY Georgiana Marin 1 Abstract In terms of data processing, classical statistical models are restrictive; it requires hypotheses, the knowledge and experience of specialists, equations,
A Property & Casualty Insurance Predictive Modeling Process in SAS
Paper AA-02-2015 A Property & Casualty Insurance Predictive Modeling Process in SAS 1.0 ABSTRACT Mei Najim, Sedgwick Claim Management Services, Chicago, Illinois Predictive analytics has been developing
Framing Business Problems as Data Mining Problems
Framing Business Problems as Data Mining Problems Asoka Diggs Data Scientist, Intel IT January 21, 2016 Legal Notices This presentation is for informational purposes only. INTEL MAKES NO WARRANTIES, EXPRESS
Data Quality Mining: Employing Classifiers for Assuring consistent Datasets
Data Quality Mining: Employing Classifiers for Assuring consistent Datasets Fabian Grüning Carl von Ossietzky Universität Oldenburg, Germany, [email protected] Abstract: Independent
How To Cluster
Data Clustering Dec 2nd, 2013 Kyrylo Bessonov Talk outline Introduction to clustering Types of clustering Supervised Unsupervised Similarity measures Main clustering algorithms k-means Hierarchical Main
ICT Perspectives on Big Data: Well Sorted Materials
ICT Perspectives on Big Data: Well Sorted Materials 3 March 2015 Contents Introduction 1 Dendrogram 2 Tree Map 3 Heat Map 4 Raw Group Data 5 For an online, interactive version of the visualisations in
Classification of Bad Accounts in Credit Card Industry
Classification of Bad Accounts in Credit Card Industry Chengwei Yuan December 12, 2014 Introduction Risk management is critical for a credit card company to survive in such competing industry. In addition
Data mining and official statistics
Quinta Conferenza Nazionale di Statistica Data mining and official statistics Gilbert Saporta président de la Société française de statistique 5@ S Roma 15, 16, 17 novembre 2000 Palazzo dei Congressi Piazzale
Why do statisticians "hate" us?
Why do statisticians "hate" us? David Hand, Heikki Mannila, Padhraic Smyth "Data mining is the analysis of (often large) observational data sets to find unsuspected relationships and to summarize the data
Business Intelligence and Decision Support Systems
Chapter 12 Business Intelligence and Decision Support Systems Information Technology For Management 7 th Edition Turban & Volonino Based on lecture slides by L. Beaubien, Providence College John Wiley
ANALYTIC HIERARCHY PROCESS (AHP) TUTORIAL
Kardi Teknomo ANALYTIC HIERARCHY PROCESS (AHP) TUTORIAL Revoledu.com Table of Contents Analytic Hierarchy Process (AHP) Tutorial... 1 Multi Criteria Decision Making... 1 Cross Tabulation... 2 Evaluation
Measurement and Metrics Fundamentals. SE 350 Software Process & Product Quality
Measurement and Metrics Fundamentals Lecture Objectives Provide some basic concepts of metrics Quality attribute metrics and measurements Reliability, validity, error Correlation and causation Discuss
A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data
A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data Athanasius Zakhary, Neamat El Gayar Faculty of Computers and Information Cairo University, Giza, Egypt
How To Use Data Mining For Loyalty Based Management
Data Mining for Loyalty Based Management Petra Hunziker, Andreas Maier, Alex Nippe, Markus Tresch, Douglas Weers, Peter Zemp Credit Suisse P.O. Box 100, CH - 8070 Zurich, Switzerland [email protected],
Analyzing Research Data Using Excel
Analyzing Research Data Using Excel Fraser Health Authority, 2012 The Fraser Health Authority ( FH ) authorizes the use, reproduction and/or modification of this publication for purposes other than commercial
