MAXIMIZING RETURN ON DIRECT MARKETING CAMPAIGNS
|
|
|
- David Chandler
- 10 years ago
- Views:
Transcription
1 MAXIMIZING RETURN ON DIRET MARKETING AMPAIGNS IN OMMERIAL BANKING S 229 Project: Final Report Oleksandra Onosova INTRODUTION Recent innovations in cloud computing and unified communications have made a technologically advanced call center a more viable option for marketing campaigns than ever. A well-executed direct advertising campaign can provide a positive return on investment by motivating customers to respond to a call-to-action. One of the preparatory stages ensuring a high return on investment consists of carefully querying a marketing database to generate a target list of respondents. In this paper, a number of machine learning techniques were used to optimize the target list of consumers of a campaign. The data for the analysis (S. Moro R. L., 20) was provided by a Portuguese banking institution seeking to sell the subscription to a bank deposit through direct calls. A prior research paper (S. Moro R. L., 20) describes the analysis performed utilizing the same data. SVM methodology was used and call duration was recognized as the most relevant feature for classification. To enhance the findings of the prior paper, call duration was excluded from the feature list and the objective was reformulated. The new goal was to select a target audience for the campaign by using machine learning techniques prior to making calls in order to maximize the net profit. DATA PREPARATION The original data consist of 6 mixed (categorical and numeric) features with no missing values. They fall into three main groups: The demographical information (age, education, job, marital status) Information related to banking(balance, prior defaults, loans) Data about the current and prior campaigns (date of the call, days passed since last contact, outcome of the prior campaign for that client, etc.) As a part of data analysis, first, additional features were added: the year feature (data description characterized it as being organized by date in an ascending order) day of the week Next, the data were scrubbed. For example, the value of the variable pdays, reflecting how many days before the current call, the customers had been contacted contained the value of - signifying that the customer was never contacted before. It was substituted by a value slightly higher than the maximum value of that feature. As the next step, the data were binned separately for various classification approaches. Binning was performed according to the distribution of the variables. For the variables that were almost uniformly distributed (age, pdays ), an equally spaced binning was used, with the exception of the tail values that were binned together. For the approximately logarithmically distributed variables balance, campaign, (where the latter signifies the number of contacts performed during this campaign and for this client), the log transformation was applied and then the
2 normal distribution was used for binning. The categorical variables were coded with levels for Naïve Bayes classifier and as matrices of 0s and s for SVM and regression classifiers. Finally, the data were split into 80% for training and 20% for testing purposes. The training dataset was further analyzed using 3-fold validation and the final results (presented below) were generated on the testing set. METHODS The following machine learning approaches for performing binary classification were attempted: Naïve Bayes, Logistic and, and SVM. The algorithms split a set of potential customers into two classes: the ones to call (i.e. those that have a sufficiently high probability of subscription) and those not to call (i.e. those that are unlikely to make a purchase). The net profit was the metric used in optimizing classifier performance and was formulated as follows: 2 Where: n number of customers in the list ustomers belonging to + ( - ) class customers that were evaluated as likely (unlikely) to make a purchase R Expected revenue from a successful call; ost of a successful call; 2 ost of a failed call, where on average, 2 < due to shorter call duration. An important additional task was adjusting the selected machine learning approaches to account for the asymmetry of the misclassification cost, i.e. taking into the account the ratio of the cost of a false positive and the cost of a false negative. Due to the disproportionately low cost of direct marketing calls in comparison with the high expected revenue in case of the successful outcome of the call, false positives were considerably cheaper than false negatives. onsequently, two techniques: ) thresholding (H. Drucker, 999) and 2) enhancing the SVM objective function to incorporate two cost parameters and then tuning all the parameters (cf. Osuna, Freund, & Girosi, 997) were attempted to implement the asymmetric cost adjustment. Thresholding consisted of finding an optimal threshold probability of а successful call. It helped identify the maximum probability of а successful sell that would make the call worthwhile. A range of thresholds from 0% to 00% was generated and the profit metric was calculated for each threshold in the range. The procedure was repeated for the three subsets of the training data as 3-fold cross-validation was used. The results were then applied to the testing set to estimate the resulting profit. The following thresholding approaches were used for various classifiers: For Naïve Bayes, and Logistic Regression the optimal threshold was calculated as follows: Where: Vi - ross-validation set i out of 3 sets 2 argmax For performing Naïve Bayes and Logistic/Probit regression, the MATLAB inbuilt functions were utilized. For SVM, the optimal threshold was tuned along with the rest of the parameters for Linear SVM and SVM with RBF Kernel: (Linear),, argmax,,
3 ,,, argmax,,, (Gaussian kernel, where W is a weight coefficient on cost of the - class) 00% RO $30,000 Thresholding Effect TPR 80% $20,000 60% 40% $0,000 20% Naïve Bayes Naïve Bayes $0 0% 0% 20% 40% 60% 80% 00% 0% 00% FPR % ustomers to be alled Fig. Fig. 2 The RO curve was generated on the test set by varying the threshold parameter and estimating the True Positive Rate and the False Positive Rate at each threshold level. The results showed that was clearly superior with respect of Naïve Bayes classifier, and most likely better that the other approaches used. The results on Figure 2 were generated by the same procedure of varying threshold level and on the same data as the RO curves, but additionally, the Profit and % ustomers to call metrics were estimated. The plots are intuitive in showing that, as the probability grows from 0% to about 50%, and, correspondingly, a higher percent of respondents is called, the net profit is rising: a slightly higher cost of the campaign is still lower than the marginal increase in the revenue. After the saturation point is reached (40%-50% probability threshold for various classifiers), the marginal cost of additional calls become higher than the expected revenue, as the probability of achieving a successful sell is lower. Only Probit regression result with is shown on the figure due to very similar results for Logit and Probit. Since the goal was to maximize the profit by varying the threshold, the ultimate metric for the classification was: max Where: t Threshold; P(S) Predicted probability of successful call 2 The second technique for adjusting the classifiers to account for the asymmetric cost function was enhancing the SVM objective function and tuning the parameters in the SVM algorithm. The objective function with two cost parameters was formulated as follows: min,, 2 ξ ξ s.t. ξ 0,,, ξ ω x b, i,, n where is the cost for the + cases, corresponding to y= W is the cost for the - cases, corresponding to y=- $Profit With the above objective function, the SVM parameters were tuned using a grid-search and maximizing the targeted profit. The algorithm is implemented with the inbuilt features of LibSVM (hih-hung hang). 3
4 In case of linear SVM, the optimal parameters and W were found with a grid search on ranges 2-3 to 2 2 and 2-3 to 2 6 respectively. The resulting maximum profit as a function of and W is shown on Figure 3. It is easily noticeable that for different values of cost, the maximum is achieved near W =8. More localized grid search resulted in =0.25 and W =0 as parameters maximizing cross-validation profit for linear SVM. For RBF kernel SVM, the optimal parameters were found in two steps. First, by searching for optimal and W (while keeping γ default value for LIBSVM), it was found that W =0 is the optimal value for all values of considered ( varied from 2-3 to 2 8 and W from 2 3 to 2 6 ), as shown on Figure 4. The value W =0 was chosen for the second grid search step as the optimal V (ross- Validation) profit was found by varying and γ ( varied from 2-3 to 2 4 and γ - from 2-9 to 2 2 ). The optimal parameters for RBF SVM model were found to be the following: =0.25, γ=0.25, and W =0. Grid Search Max V Profit over, W x 0 5 x 0 5 Grid Search Max Profit: Linear SVM W, W 0-2 Fig. 3 Grid Search for Max Profit RBF SVM.2 x Fig γ W Fig. 5 QUANTIFIATION OF THE OST MATRIX The average cost of minute of a marketing call, performed by a call center is approximately $; the average call duration for a successful call is 0 minutes, whereas the call duration for a failed marketing call is 4 minutes ("3 Reasons the all enter Is Far From Dead", 202). The costs used for the analysis were respectively = $0 and 2 = $4. Lifetime value for an average customer relationship in retail banking is $50 (Milman, 202). The value of the customer subscribing to the deposit account was assumed to be $60, comprising approximately 40% of the total value of a customer relationship. Thus, the net profit of a customer subscription is ($60 - $0 = $50). The combined net profit matrix becomes: 4 Unsuccessful call Successful all all is not made 0 0 all is made -$4 $50
5 RESULTS AND ONLUSION SVM was found to produce the best optimization in terms of the profit margin. In particular, SVM with the Gaussian Kernel and the parameters tuned in the grid search illustrated above was found to be the champion classifier among the attempted ones. The chart on the right (blue bars) shows the profit on the test set, Base ase generated by applying the models $- $5,000 $0,000 $5,000 $20,000 $25,000 $30,000 with thresholds (and other parameters in the case of SVM) found to maximize on the training Test Profit Diff from Oracle View set. The base case used for comparison shows the profit generated by calling 00% of the customers on the test set list. The red bars show the Threshold Oracle View estimated by using the optimal threshold with respect to the test set. The optimal threshold was found by grid search, with the goal of maximizing the test set profit. The minimal difference between the Test Profit and the Threshold Oracle View demonstrates robustness of threshold estimation on the training set. In conclusion, using SVM with Gaussian kernel resulted in a 4% increase in marketing profit of the campaign. In the given dataset, the number of successful sells was uncharacteristically high about a tenth of the calls. In the more typical marketing campaigns, the successful sells comprise an even smaller percentage of calls, leading to higher costs. In that case, the asymmetry of the cost function would be even more pronounced and thus the benefit of the analysis would be even higher. In addition to maximizing profit, other goals such as minimizing customer frustration could be achieved by trimming the target customer list. Future direction for quantifying the effect of the campaign would be measuring the negative impact of unsuccessful marketing calls and assigning a negative cost to them to further diminish the number of calls to the customers. That would effectively change the cost function and leave the underlying problem unchanged. In addition, a more thorough grid-search of SVM parameters could be attempted, with all RBF parameters being varied along all the dimensions at the same time, if time permits. Finally, other classification techniques that are known to effectively address the asymmetric cost function could be attempted, such as Ada Boost algorithm. REFERENES Naïve Bayes Total Profiton Test Set vs Treshold Oracle Model View "3 Reasons the all enter Is Far From Dead". (202, April 24). Retrieved November 0, 202, from Mashable: hih-hung hang,.-j. L. (n.d.). "LIBSVM -- A Library for Support Vector Machines". Retrieved from H. Drucker, D. W. (999). "Support vector machines for spam classification". (p. 0(5)). IEEE. Milman, S. (202). "Bank and redit Union Business Strategy and ustomer Life Time Value". Retrieved November 0, 202, from Optirate: S. Moro, R. L. (20, October). Bank Marketing Data Set. Retrieved November 0, 202, from U Irvine Machine Learning Repository: S. Moro, R. L. (20). Using Data Mining for Bank Direct Marketing: An Application of the RISP-DM Methodology. Proceedings of the European Simulation and Modelling onference - ESM'20 (pp. pp. 7-2). Guimarães, Portugal: EUROSIS. 5
UNDERSTANDING THE EFFECTIVENESS OF BANK DIRECT MARKETING Tarun Gupta, Tong Xia and Diana Lee
UNDERSTANDING THE EFFECTIVENESS OF BANK DIRECT MARKETING Tarun Gupta, Tong Xia and Diana Lee 1. Introduction There are two main approaches for companies to promote their products / services: through mass
Data Mining Application in Direct Marketing: Identifying Hot Prospects for Banking Product
Data Mining Application in Direct Marketing: Identifying Hot Prospects for Banking Product Sagarika Prusty Web Data Mining (ECT 584),Spring 2013 DePaul University,Chicago [email protected] Keywords:
Comparing the Results of Support Vector Machines with Traditional Data Mining Algorithms
Comparing the Results of Support Vector Machines with Traditional Data Mining Algorithms Scott Pion and Lutz Hamel Abstract This paper presents the results of a series of analyses performed on direct mail
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Role of Customer Response Models in Customer Solicitation Center s Direct Marketing Campaign
Role of Customer Response Models in Customer Solicitation Center s Direct Marketing Campaign Arun K Mandapaka, Amit Singh Kushwah, Dr.Goutam Chakraborty Oklahoma State University, OK, USA ABSTRACT Direct
Search Taxonomy. Web Search. Search Engine Optimization. Information Retrieval
Information Retrieval INFO 4300 / CS 4300! Retrieval models Older models» Boolean retrieval» Vector Space model Probabilistic Models» BM25» Language models Web search» Learning to Rank Search Taxonomy!
Making Sense of the Mayhem: Machine Learning and March Madness
Making Sense of the Mayhem: Machine Learning and March Madness Alex Tran and Adam Ginzberg Stanford University [email protected] [email protected] I. Introduction III. Model The goal of our research
Scalable Developments for Big Data Analytics in Remote Sensing
Scalable Developments for Big Data Analytics in Remote Sensing Federated Systems and Data Division Research Group High Productivity Data Processing Dr.-Ing. Morris Riedel et al. Research Group Leader,
Predict the Popularity of YouTube Videos Using Early View Data
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
The Artificial Prediction Market
The Artificial Prediction Market Adrian Barbu Department of Statistics Florida State University Joint work with Nathan Lay, Siemens Corporate Research 1 Overview Main Contributions A mathematical theory
Data Mining. Nonlinear Classification
Data Mining Unit # 6 Sajjad Haider Fall 2014 1 Nonlinear Classification Classes may not be separable by a linear boundary Suppose we randomly generate a data set as follows: X has range between 0 to 15
Predict Influencers in the Social Network
Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, [email protected] Department of Electrical Engineering, Stanford University Abstract Given two persons
Active Learning SVM for Blogs recommendation
Active Learning SVM for Blogs recommendation Xin Guan Computer Science, George Mason University Ⅰ.Introduction In the DH Now website, they try to review a big amount of blogs and articles and find the
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Applied Data Mining Analysis: A Step-by-Step Introduction Using Real-World Data Sets
Applied Data Mining Analysis: A Step-by-Step Introduction Using Real-World Data Sets http://info.salford-systems.com/jsm-2015-ctw August 2015 Salford Systems Course Outline Demonstration of two classification
Evaluation and Comparison of Data Mining Techniques Over Bank Direct Marketing
Evaluation and Comparison of Data Mining Techniques Over Bank Direct Marketing Niharika Sharma 1, Arvinder Kaur 2, Sheetal Gandotra 3, Dr Bhawna Sharma 4 B.E. Final Year Student, Department of Computer
Support Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France [email protected] Massimiliano
USING DATA MINING FOR BANK DIRECT MARKETING: AN APPLICATION OF THE CRISP-DM METHODOLOGY
USING DATA MINING FOR BANK DIRECT MARKETING: AN APPLICATION OF THE CRISP-DM METHODOLOGY Sérgio Moro and Raul M. S. Laureano Instituto Universitário de Lisboa (ISCTE IUL) Av.ª das Forças Armadas 1649-026
Machine Learning Logistic Regression
Machine Learning Logistic Regression Jeff Howbert Introduction to Machine Learning Winter 2012 1 Logistic regression Name is somewhat misleading. Really a technique for classification, not regression.
APPLICATION OF DATA MINING TECHNIQUES FOR DIRECT MARKETING. Anatoli Nachev
86 ITHEA APPLICATION OF DATA MINING TECHNIQUES FOR DIRECT MARKETING Anatoli Nachev Abstract: This paper presents a case study of data mining modeling techniques for direct marketing. It focuses to three
Data Mining - Evaluation of Classifiers
Data Mining - Evaluation of Classifiers Lecturer: JERZY STEFANOWSKI Institute of Computing Sciences Poznan University of Technology Poznan, Poland Lecture 4 SE Master Course 2008/2009 revised for 2010
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
Machine Learning in Spam Filtering
Machine Learning in Spam Filtering A Crash Course in ML Konstantin Tretyakov [email protected] Institute of Computer Science, University of Tartu Overview Spam is Evil ML for Spam Filtering: General Idea, Problems.
MERGING BUSINESS KPIs WITH PREDICTIVE MODEL KPIs FOR BINARY CLASSIFICATION MODEL SELECTION
MERGING BUSINESS KPIs WITH PREDICTIVE MODEL KPIs FOR BINARY CLASSIFICATION MODEL SELECTION Matthew A. Lanham & Ralph D. Badinelli Virginia Polytechnic Institute and State University Department of Business
Introduction to Logistic Regression
OpenStax-CNX module: m42090 1 Introduction to Logistic Regression Dan Calderon This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 3.0 Abstract Gives introduction
STATISTICA. Financial Institutions. Case Study: Credit Scoring. and
Financial Institutions and STATISTICA Case Study: Credit Scoring STATISTICA Solutions for Business Intelligence, Data Mining, Quality Control, and Web-based Analytics Table of Contents INTRODUCTION: WHAT
Azure Machine Learning, SQL Data Mining and R
Azure Machine Learning, SQL Data Mining and R Day-by-day Agenda Prerequisites No formal prerequisites. Basic knowledge of SQL Server Data Tools, Excel and any analytical experience helps. Best of all:
ECLT5810 E-Commerce Data Mining Technique SAS Enterprise Miner -- Regression Model I. Regression Node
Enterprise Miner - Regression 1 ECLT5810 E-Commerce Data Mining Technique SAS Enterprise Miner -- Regression Model I. Regression Node 1. Some background: Linear attempts to predict the value of a continuous
A Data-Driven Approach to Predict the. Success of Bank Telemarketing
A Data-Driven Approach to Predict the Success of Bank Telemarketing Sérgio Moro a, Paulo Cortez b Paulo Rita a a ISCTE - University Institute of Lisbon, 1649-026 Lisboa, Portugal b ALGORITMI Research Centre,
Employer Health Insurance Premium Prediction Elliott Lui
Employer Health Insurance Premium Prediction Elliott Lui 1 Introduction The US spends 15.2% of its GDP on health care, more than any other country, and the cost of health insurance is rising faster than
SVM Ensemble Model for Investment Prediction
19 SVM Ensemble Model for Investment Prediction Chandra J, Assistant Professor, Department of Computer Science, Christ University, Bangalore Siji T. Mathew, Research Scholar, Christ University, Dept of
E-commerce Transaction Anomaly Classification
E-commerce Transaction Anomaly Classification Minyong Lee [email protected] Seunghee Ham [email protected] Qiyi Jiang [email protected] I. INTRODUCTION Due to the increasing popularity of e-commerce
BIDM Project. Predicting the contract type for IT/ITES outsourcing contracts
BIDM Project Predicting the contract type for IT/ITES outsourcing contracts N a n d i n i G o v i n d a r a j a n ( 6 1 2 1 0 5 5 6 ) The authors believe that data modelling can be used to predict if an
1 Topic. 2 Scilab. 2.1 What is Scilab?
1 Topic Data Mining with Scilab. I know the name "Scilab" for a long time (http://www.scilab.org/en). For me, it is a tool for numerical analysis. It seemed not interesting in the context of the statistical
Early defect identification of semiconductor processes using machine learning
STANFORD UNIVERISTY MACHINE LEARNING CS229 Early defect identification of semiconductor processes using machine learning Friday, December 16, 2011 Authors: Saul ROSA Anton VLADIMIROV Professor: Dr. Andrew
MACHINE LEARNING IN HIGH ENERGY PHYSICS
MACHINE LEARNING IN HIGH ENERGY PHYSICS LECTURE #1 Alex Rogozhnikov, 2015 INTRO NOTES 4 days two lectures, two practice seminars every day this is introductory track to machine learning kaggle competition!
Knowledge Discovery from patents using KMX Text Analytics
Knowledge Discovery from patents using KMX Text Analytics Dr. Anton Heijs [email protected] Treparel Abstract In this white paper we discuss how the KMX technology of Treparel can help searchers
Response variables assume only two values, say Y j = 1 or = 0, called success and failure (spam detection, credit scoring, contracting.
Prof. Dr. J. Franke All of Statistics 1.52 Binary response variables - logistic regression Response variables assume only two values, say Y j = 1 or = 0, called success and failure (spam detection, credit
Performance Analysis of Naive Bayes and J48 Classification Algorithm for Data Classification
Performance Analysis of Naive Bayes and J48 Classification Algorithm for Data Classification Tina R. Patil, Mrs. S. S. Sherekar Sant Gadgebaba Amravati University, Amravati [email protected], [email protected]
Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: [email protected] Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
Server Load Prediction
Server Load Prediction Suthee Chaidaroon ([email protected]) Joon Yeong Kim ([email protected]) Jonghan Seo ([email protected]) Abstract Estimating server load average is one of the methods that
Beating the MLB Moneyline
Beating the MLB Moneyline Leland Chen [email protected] Andrew He [email protected] 1 Abstract Sports forecasting is a challenging task that has similarities to stock market prediction, requiring time-series
Tweaking Naïve Bayes classifier for intelligent spam detection
682 Tweaking Naïve Bayes classifier for intelligent spam detection Ankita Raturi 1 and Sunil Pranit Lal 2 1 University of California, Irvine, CA 92697, USA. [email protected] 2 School of Computing, Information
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! [email protected]! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
Logistic Regression for Spam Filtering
Logistic Regression for Spam Filtering Nikhila Arkalgud February 14, 28 Abstract The goal of the spam filtering problem is to identify an email as a spam or not spam. One of the classic techniques used
Practical Data Science with Azure Machine Learning, SQL Data Mining, and R
Practical Data Science with Azure Machine Learning, SQL Data Mining, and R Overview This 4-day class is the first of the two data science courses taught by Rafal Lukawiecki. Some of the topics will be
Classification and Regression by randomforest
Vol. 2/3, December 02 18 Classification and Regression by randomforest Andy Liaw and Matthew Wiener Introduction Recently there has been a lot of interest in ensemble learning methods that generate many
WebFOCUS RStat. RStat. Predict the Future and Make Effective Decisions Today. WebFOCUS RStat
Information Builders enables agile information solutions with business intelligence (BI) and integration technologies. WebFOCUS the most widely utilized business intelligence platform connects to any enterprise
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
SURVIVABILITY OF COMPLEX SYSTEM SUPPORT VECTOR MACHINE BASED APPROACH
1 SURVIVABILITY OF COMPLEX SYSTEM SUPPORT VECTOR MACHINE BASED APPROACH Y, HONG, N. GAUTAM, S. R. T. KUMARA, A. SURANA, H. GUPTA, S. LEE, V. NARAYANAN, H. THADAKAMALLA The Dept. of Industrial Engineering,
Predicting Flight Delays
Predicting Flight Delays Dieterich Lawson [email protected] William Castillo [email protected] Introduction Every year approximately 20% of airline flights are delayed or cancelled, costing
Strategic Online Advertising: Modeling Internet User Behavior with
2 Strategic Online Advertising: Modeling Internet User Behavior with Patrick Johnston, Nicholas Kristoff, Heather McGinness, Phuong Vu, Nathaniel Wong, Jason Wright with William T. Scherer and Matthew
Predictive Data modeling for health care: Comparative performance study of different prediction models
Predictive Data modeling for health care: Comparative performance study of different prediction models Shivanand Hiremath [email protected] National Institute of Industrial Engineering (NITIE) Vihar
Lasso-based Spam Filtering with Chinese Emails
Journal of Computational Information Systems 8: 8 (2012) 3315 3322 Available at http://www.jofcis.com Lasso-based Spam Filtering with Chinese Emails Zunxiong LIU 1, Xianlong ZHANG 1,, Shujuan ZHENG 2 1
Data Mining Techniques for Prognosis in Pancreatic Cancer
Data Mining Techniques for Prognosis in Pancreatic Cancer by Stuart Floyd A Thesis Submitted to the Faculty of the WORCESTER POLYTECHNIC INSTITUE In partial fulfillment of the requirements for the Degree
Anti-Spam Filter Based on Naïve Bayes, SVM, and KNN model
AI TERM PROJECT GROUP 14 1 Anti-Spam Filter Based on,, and model Yun-Nung Chen, Che-An Lu, Chao-Yu Huang Abstract spam email filters are a well-known and powerful type of filters. We construct different
KEITH LEHNERT AND ERIC FRIEDRICH
MACHINE LEARNING CLASSIFICATION OF MALICIOUS NETWORK TRAFFIC KEITH LEHNERT AND ERIC FRIEDRICH 1. Introduction 1.1. Intrusion Detection Systems. In our society, information systems are everywhere. They
On the Path to an Ideal ROC Curve: Considering Cost Asymmetry in Learning Classifiers
On the Path to an Ideal ROC Curve: Considering Cost Asymmetry in Learning Classifiers Francis R. Bach Computer Science Division University of California Berkeley, CA 9472 [email protected] Abstract
Class #6: Non-linear classification. ML4Bio 2012 February 17 th, 2012 Quaid Morris
Class #6: Non-linear classification ML4Bio 2012 February 17 th, 2012 Quaid Morris 1 Module #: Title of Module 2 Review Overview Linear separability Non-linear classification Linear Support Vector Machines
Beating the NCAA Football Point Spread
Beating the NCAA Football Point Spread Brian Liu Mathematical & Computational Sciences Stanford University Patrick Lai Computer Science Department Stanford University December 10, 2010 1 Introduction Over
BIOINF 585 Fall 2015 Machine Learning for Systems Biology & Clinical Informatics http://www.ccmb.med.umich.edu/node/1376
Course Director: Dr. Kayvan Najarian (DCM&B, [email protected]) Lectures: Labs: Mondays and Wednesdays 9:00 AM -10:30 AM Rm. 2065 Palmer Commons Bldg. Wednesdays 10:30 AM 11:30 AM (alternate weeks) Rm.
Decision Support Systems
Decision Support Systems 50 (2011) 602 613 Contents lists available at ScienceDirect Decision Support Systems journal homepage: www.elsevier.com/locate/dss Data mining for credit card fraud: A comparative
Data Mining Classification: Decision Trees
Data Mining Classification: Decision Trees Classification Decision Trees: what they are and how they work Hunt s (TDIDT) algorithm How to select the best split How to handle Inconsistent data Continuous
Machine Learning Final Project Spam Email Filtering
Machine Learning Final Project Spam Email Filtering March 2013 Shahar Yifrah Guy Lev Table of Content 1. OVERVIEW... 3 2. DATASET... 3 2.1 SOURCE... 3 2.2 CREATION OF TRAINING AND TEST SETS... 4 2.3 FEATURE
Statistical Data Mining. Practical Assignment 3 Discriminant Analysis and Decision Trees
Statistical Data Mining Practical Assignment 3 Discriminant Analysis and Decision Trees In this practical we discuss linear and quadratic discriminant analysis and tree-based classification techniques.
Support Vector Machines Explained
March 1, 2009 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),
Microsoft Azure Machine learning Algorithms
Microsoft Azure Machine learning Algorithms Tomaž KAŠTRUN @tomaz_tsql [email protected] http://tomaztsql.wordpress.com Our Sponsors Speaker info https://tomaztsql.wordpress.com Agenda Focus on explanation
In this tutorial, we try to build a roc curve from a logistic regression.
Subject In this tutorial, we try to build a roc curve from a logistic regression. Regardless the software we used, even for commercial software, we have to prepare the following steps when we want build
Keywords : Data Warehouse, Data Warehouse Testing, Lifecycle based Testing
Volume 4, Issue 12, December 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com A Lifecycle
Data Mining - The Next Mining Boom?
Howard Ong Principal Consultant Aurora Consulting Pty Ltd Abstract This paper introduces Data Mining to its audience by explaining Data Mining in the context of Corporate and Business Intelligence Reporting.
Getting Even More Out of Ensemble Selection
Getting Even More Out of Ensemble Selection Quan Sun Department of Computer Science The University of Waikato Hamilton, New Zealand [email protected] ABSTRACT Ensemble Selection uses forward stepwise
Drugs store sales forecast using Machine Learning
Drugs store sales forecast using Machine Learning Hongyu Xiong (hxiong2), Xi Wu (wuxi), Jingying Yue (jingying) 1 Introduction Nowadays medical-related sales prediction is of great interest; with reliable
Consistent Binary Classification with Generalized Performance Metrics
Consistent Binary Classification with Generalized Performance Metrics Nagarajan Natarajan Joint work with Oluwasanmi Koyejo, Pradeep Ravikumar and Inderjit Dhillon UT Austin Nov 4, 2014 Problem and Motivation
A Basic Guide to Modeling Techniques for All Direct Marketing Challenges
A Basic Guide to Modeling Techniques for All Direct Marketing Challenges Allison Cornia Database Marketing Manager Microsoft Corporation C. Olivia Rud Executive Vice President Data Square, LLC Overview
Analysis Tools and Libraries for BigData
+ Analysis Tools and Libraries for BigData Lecture 02 Abhijit Bendale + Office Hours 2 n Terry Boult (Waiting to Confirm) n Abhijit Bendale (Tue 2:45 to 4:45 pm). Best if you email me in advance, but I
Towards better accuracy for Spam predictions
Towards better accuracy for Spam predictions Chengyan Zhao Department of Computer Science University of Toronto Toronto, Ontario, Canada M5S 2E4 [email protected] Abstract Spam identification is crucial
How To Cluster
Data Clustering Dec 2nd, 2013 Kyrylo Bessonov Talk outline Introduction to clustering Types of clustering Supervised Unsupervised Similarity measures Main clustering algorithms k-means Hierarchical Main
Classification Problems
Classification Read Chapter 4 in the text by Bishop, except omit Sections 4.1.6, 4.1.7, 4.2.4, 4.3.3, 4.3.5, 4.3.6, 4.4, and 4.5. Also, review sections 1.5.1, 1.5.2, 1.5.3, and 1.5.4. Classification Problems
Benchmarking of different classes of models used for credit scoring
Benchmarking of different classes of models used for credit scoring We use this competition as an opportunity to compare the performance of different classes of predictive models. In particular we want
Music Mood Classification
Music Mood Classification CS 229 Project Report Jose Padial Ashish Goel Introduction The aim of the project was to develop a music mood classifier. There are many categories of mood into which songs may
Big Data Analytics CSCI 4030
High dim. data Graph data Infinite data Machine learning Apps Locality sensitive hashing PageRank, SimRank Filtering data streams SVM Recommen der systems Clustering Community Detection Web advertising
Acknowledgments. Data Mining with Regression. Data Mining Context. Overview. Colleagues
Data Mining with Regression Teaching an old dog some new tricks Acknowledgments Colleagues Dean Foster in Statistics Lyle Ungar in Computer Science Bob Stine Department of Statistics The School of the
Lecture 6: Logistic Regression
Lecture 6: CS 194-10, Fall 2011 Laurent El Ghaoui EECS Department UC Berkeley September 13, 2011 Outline Outline Classification task Data : X = [x 1,..., x m]: a n m matrix of data points in R n. y { 1,
Decision Support System For A Customer Relationship Management Case Study
61 Decision Support System For A Customer Relationship Management Case Study Ozge Kart 1, Alp Kut 1, and Vladimir Radevski 2 1 Dokuz Eylul University, Izmir, Turkey {ozge, alp}@cs.deu.edu.tr 2 SEE University,
Active Learning with Boosting for Spam Detection
Active Learning with Boosting for Spam Detection Nikhila Arkalgud Last update: March 22, 2008 Active Learning with Boosting for Spam Detection Last update: March 22, 2008 1 / 38 Outline 1 Spam Filters
Car Insurance. Prvák, Tomi, Havri
Car Insurance Prvák, Tomi, Havri Sumo report - expectations Sumo report - reality Bc. Jan Tomášek Deeper look into data set Column approach Reminder What the hell is this competition about??? Attributes
Support Vector Machine (SVM)
Support Vector Machine (SVM) CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
How To Identify A Churner
2012 45th Hawaii International Conference on System Sciences A New Ensemble Model for Efficient Churn Prediction in Mobile Telecommunication Namhyoung Kim, Jaewook Lee Department of Industrial and Management
Machine learning for algo trading
Machine learning for algo trading An introduction for nonmathematicians Dr. Aly Kassam Overview High level introduction to machine learning A machine learning bestiary What has all this got to do with
STATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
Logistic Regression (1/24/13)
STA63/CBB540: Statistical methods in computational biology Logistic Regression (/24/3) Lecturer: Barbara Engelhardt Scribe: Dinesh Manandhar Introduction Logistic regression is model for regression used
CS 229, Autumn 2011 Modeling the Stock Market Using Twitter Sentiment Analysis
CS 229, Autumn 2011 Modeling the Stock Market Using Twitter Sentiment Analysis Team members: Daniel Debbini, Philippe Estin, Maxime Goutagny Supervisor: Mihai Surdeanu (with John Bauer) 1 Introduction
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
ElegantJ BI. White Paper. The Competitive Advantage of Business Intelligence (BI) Forecasting and Predictive Analysis
ElegantJ BI White Paper The Competitive Advantage of Business Intelligence (BI) Forecasting and Predictive Analysis Integrated Business Intelligence and Reporting for Performance Management, Operational
