Car Insurance. Havránek, Pokorný, Tomášek

Similar documents
Car Insurance. Prvák, Tomi, Havri

MHI3000 Big Data Analytics for Health Care Final Project Report

Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence

Designing a neural network for forecasting financial time series

Predict Influencers in the Social Network

Data Mining. Nonlinear Classification

Data Mining Practical Machine Learning Tools and Techniques

Supervised Learning (Big Data Analytics)

Data Mining Algorithms Part 1. Dejan Sarka

Decision Trees from large Databases: SLIQ

Neural Networks and Support Vector Machines

Employer Health Insurance Premium Prediction Elliott Lui

Model Combination. 24 Novembre 2009

Classification algorithm in Data mining: An Overview

LCs for Binary Classification

Predictive Data modeling for health care: Comparative performance study of different prediction models

Question 2 Naïve Bayes (16 points)

CSE 473: Artificial Intelligence Autumn 2010

Neural Networks for Sentiment Detection in Financial Text

Ensemble Methods. Knowledge Discovery and Data Mining 2 (VU) ( ) Roman Kern. KTI, TU Graz

Chapter 6. The stacking ensemble approach

Detecting Corporate Fraud: An Application of Machine Learning

A Content based Spam Filtering Using Optical Back Propagation Technique

Making Sense of the Mayhem: Machine Learning and March Madness

Microsoft Azure Machine learning Algorithms

Machine Learning Big Data using Map Reduce

Comparing the Results of Support Vector Machines with Traditional Data Mining Algorithms

Data Mining Lab 5: Introduction to Neural Networks

Data Mining Techniques for Prognosis in Pancreatic Cancer

Machine Learning Final Project Spam Filtering

Search Taxonomy. Web Search. Search Engine Optimization. Information Retrieval

The Data Mining Process

SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS

High-Performance Signature Recognition Method using SVM

Machine learning algorithms for predicting roadside fine particulate matter concentration level in Hong Kong Central

L25: Ensemble learning

Predicting Good Probabilities With Supervised Learning

Leveraging Ensemble Models in SAS Enterprise Miner

An Introduction to Data Mining

STAT 503X Case Study 2: Italian Olive Oils

Application of Event Based Decision Tree and Ensemble of Data Driven Methods for Maintenance Action Recommendation

Event driven trading new studies on innovative way. of trading in Forex market. Michał Osmoła INIME live 23 February 2016

Feature Subset Selection in Spam Detection

Machine Learning Logistic Regression

Big Data Analytics CSCI 4030

Beating the MLB Moneyline

NEURAL NETWORKS A Comprehensive Foundation

Lecture 6. Artificial Neural Networks

Chapter 4: Artificial Neural Networks

Predicting Flight Delays

Stock Portfolio Selection using Data Mining Approach

Electroencephalography Analysis Using Neural Network and Support Vector Machine during Sleep

Machine learning for algo trading

Knowledge Discovery and Data Mining

Data Mining: An Overview. David Madigan

How To Predict The Outcome Of The Ncaa Basketball Tournament

Data Mining. Supervised Methods. Ciro Donalek Ay/Bi 199ab: Methods of Sciences hcp://esci101.blogspot.

Active Learning SVM for Blogs recommendation

White Paper. Redefine Your Analytics Journey With Self-Service Data Discovery and Interactive Predictive Analytics

HYBRID PROBABILITY BASED ENSEMBLES FOR BANKRUPTCY PREDICTION

Machine Learning Algorithms and Predictive Models for Undergraduate Student Retention

Runtime Hardware Reconfiguration using Machine Learning

Data Mining - Evaluation of Classifiers

How To Make A Credit Risk Model For A Bank Account

Sense Making in an IOT World: Sensor Data Analysis with Deep Learning

How to Win at the Track

6.2.8 Neural networks for data mining

Why Ensembles Win Data Mining Competitions

BIDM Project. Predicting the contract type for IT/ITES outsourcing contracts

BIOINF 585 Fall 2015 Machine Learning for Systems Biology & Clinical Informatics

A Feature Selection Methodology for Steganalysis

Learning from Diversity

CI6227: Data Mining. Lesson 11b: Ensemble Learning. Data Analytics Department, Institute for Infocomm Research, A*STAR, Singapore.

Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski

Comparison of Bagging, Boosting and Stacking Ensembles Applied to Real Estate Appraisal

Proactive Drive Failure Prediction for Large Scale Storage Systems

Knowledge Discovery and Data Mining. Bootstrap review. Bagging Important Concepts. Notes. Lecture 19 - Bagging. Tom Kelsey. Notes

FRAUD DETECTION IN ELECTRIC POWER DISTRIBUTION NETWORKS USING AN ANN-BASED KNOWLEDGE-DISCOVERY PROCESS

Machine Learning in Python with scikit-learn. O Reilly Webcast Aug. 2014

Programming Exercise 3: Multi-class Classification and Neural Networks

Chapter 6: Episode discovery process

The Artificial Prediction Market

Advanced analytics at your hands

1. Classification problems

CSCI567 Machine Learning (Fall 2014)

Journal of Asian Scientific Research COMPARISON OF THREE CLASSIFICATION ALGORITHMS FOR PREDICTING PM2.5 IN HONG KONG RURAL AREA.

Monotonicity Hints. Abstract

A fast multi-class SVM learning method for huge databases

Fast Analytics on Big Data with H20

Steven C.H. Hoi. School of Computer Engineering Nanyang Technological University Singapore

Semi-Supervised Support Vector Machines and Application to Spam Filtering

Impact of Feature Selection on the Performance of Wireless Intrusion Detection Systems

A simple application of Artificial Neural Network to cloud classification

Server Load Prediction

Knowledge Discovery and Data Mining

SVM Ensemble Model for Investment Prediction

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

Transcription:

Car Insurance Havránek, Pokorný, Tomášek

Outline Data overview Horizontal approach + Decision tree/forests Vertical (column) approach + Neural networks SVM

Data overview Customers Viewed policies Bought policy

Horizontal approach Štěpán Havránek

Horizontal approach Horizontal approach Model: decision trees Input: customer info Output: bought policy Last time results: 50-80% mean validation error per one output attribute By C4.5 alg

Horizontal approach Overfitted trees pruning Simplified the models Mean validation error decreased 30-65% per one output attribute Wider input data Customer info + one viewed policy Better results: 10-45% MVE

Horizontal approach Observation: Output attributes aren t independent New trees deciding subsets of the output attributes All the possible subsets => Out of memory exception Only manually picked subsets Singletons, pairs, 6 of 7, all Mean validation error much more better than independent products 15-80% Only observed values Running time > 12h

Horizontal approach With the subsets we have more than one tree for each output attribute Save them to the forests Voting model By validation score More progressive Less progressive Results with less wide input data (running time) More progressive Total overfit on the best tree => useless results Less progressive Variant results Not good in official rating (~ 0.2) Good for others

Column approach Bc. Jan Tomášek

Column approach + N-net Separated Neural network for each Class Neural network inputs last three entries in class as characteristic vector car age original range 0-80 cropped to 0-30 car value a - i scaled to (0-1) Levenberg-Marquardt backpropagation

Neural network - basics Oriented graph weights bias (+) strong in finding linear dependence (-) black box

Neural network 1 Neural network structure one hidden layer with 15 neurons result ~ 0.529

Neural Network 2 added more features risk factor cost result ~0.535

Neural network 3 Deep network More networks 20 networks up to 3 layers more different networks bizarre topologies result ~0.528 probably overfit

Neural network 4 Manual selection of good networks result ~0.537 total 8 networks

Neural network results Best solution(0.53745) about the same as last quoted benchmark(0.53793) nn probably learns to use last :) best solution of our group possible improvement Adaptive Boosting lear something else where last is failing combine with last quoted (how?), combine with neural network problem very fast convergence to local optimum 10 iterations almost no improvement after

scikit-learn SVC s Michal Pokorný

scikit-learn SVC s Support vector machine classifiers SVM s find the optimal separating hyperplane in a high-dimensional example space Kernel function can be interpreted as searching for similar browsing histories Basic SVM s just give a binary result, so N of them are trained for more-than-binary attributes Can be tweaked to give probabilities, not just predictions (Some simple approaches don t guarantee that.) Simplification: suggest the most likely plan the user actually looked at So, let s score every browsed plan Naive Bayes assumption : take one SVC for every attribute and multiply together their probability results

scikit-learn SVC s Tested feature ideas: Static features : day of the week, group size, homeowner, car age+value, risk factor, oldest/youngest age, married couple?, cost Histogram of browsed attribute values We saw A=0 5 times, A=1 2 times,... Most commonly browsed plan Last browsed plan, first browsed plan Unfortunately, no combinations improved upon the trivial solution :( (The best reached accuracy equals that of the benchmark model.) Overfitting? Not too likely, since changing regularization parameters of the SVC s didn t help

scikit-learn SVC s But, while the trivial solution gives 0.53739.

scikit-learn SVC s Since neither us, nor actual machine learning professionals found any significantly better solutions, the data probably just isn t there. Maybe more features could help. Website analytics could give us visit lengths, tracking of user activity, etc. - those are some ideas of useful predictors of plans or plan features the user might be interested in. Finally, please accept an apology for my physical absence - this is the result of a planning/scheduling problem concerning my student duties.

Q & A