SUCCESSFUL PREDICTION OF HORSE RACING RESULTS USING A NEURAL NETWORK



Similar documents
Chapter 4: Artificial Neural Networks

Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence

Neural Networks and Support Vector Machines

Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski

Neural Computation - Assignment

COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS

Predictive Dynamix Inc

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

Feedforward Neural Networks and Backpropagation

NEURAL NETWORKS A Comprehensive Foundation

An Introduction to Neural Networks

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 8: Multi-Layer Perceptrons

TRAINING A LIMITED-INTERCONNECT, SYNTHETIC NEURAL IC

Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승

Neural Networks and Back Propagation Algorithm

Performance Evaluation On Human Resource Management Of China S Commercial Banks Based On Improved Bp Neural Networks

Machine Learning and Data Mining -

IFT3395/6390. Machine Learning from linear regression to Neural Networks. Machine Learning. Training Set. t (3.5, -2,..., 127, 0,...

Analecta Vol. 8, No. 2 ISSN

Neural network software tool development: exploring programming language options

IBM SPSS Neural Networks 22

Automated Stellar Classification for Large Surveys with EKF and RBF Neural Networks

AUTOMATION OF ENERGY DEMAND FORECASTING. Sanzad Siddique, B.S.

Neural Networks algorithms and applications

Novelty Detection in image recognition using IRF Neural Networks properties

Neural Network Add-in

EFFICIENT DATA PRE-PROCESSING FOR DATA MINING

Role of Neural network in data mining

Neural Network Predictor for Fraud Detection: A Study Case for the Federal Patrimony Department

Neural Network Design in Cloud Computing

Predict Influencers in the Social Network

EVALUATION OF NEURAL NETWORK BASED CLASSIFICATION SYSTEMS FOR CLINICAL CANCER DATA CLASSIFICATION

The Research of Data Mining Based on Neural Networks

1. Classification problems

Digital image processing

Data Mining Techniques Chapter 7: Artificial Neural Networks

The Scientific Data Mining Process

3 An Illustrative Example

Optimum Design of Worm Gears with Multiple Computer Aided Techniques

APPLICATION OF ARTIFICIAL NEURAL NETWORKS USING HIJRI LUNAR TRANSACTION AS EXTRACTED VARIABLES TO PREDICT STOCK TREND DIRECTION

Performance Evaluation of Artificial Neural. Networks for Spatial Data Analysis

Field Data Recovery in Tidal System Using Artificial Neural Networks (ANNs)

AN APPLICATION OF TIME SERIES ANALYSIS FOR WEATHER FORECASTING

Power Prediction Analysis using Artificial Neural Network in MS Excel

degrees of freedom and are able to adapt to the task they are supposed to do [Gupta].

Self-Organising Data Mining

Comparison of K-means and Backpropagation Data Mining Algorithms

Stock Prediction using Artificial Neural Networks

Machine Learning and Pattern Recognition Logistic Regression

Comparison of Supervised and Unsupervised Learning Algorithms for Pattern Classification

Temporal Difference Learning in the Tetris Game

Data Mining and Neural Networks in Stata

Accurate and robust image superresolution by neural processing of local image representations

OPTIMUM LEARNING RATE FOR CLASSIFICATION PROBLEM

Application of Neural Network in User Authentication for Smart Home System

Open Access Research on Application of Neural Network in Computer Network Security Evaluation. Shujuan Jin *

NTC Project: S01-PH10 (formerly I01-P10) 1 Forecasting Women s Apparel Sales Using Mathematical Modeling

Data quality in Accounting Information Systems

SMORN-VII REPORT NEURAL NETWORK BENCHMARK ANALYSIS RESULTS & FOLLOW-UP 96. Özer CIFTCIOGLU Istanbul Technical University, ITU. and

Data Mining. Supervised Methods. Ciro Donalek Ay/Bi 199ab: Methods of Sciences hcp://esci101.blogspot.

Machine Learning: Multi Layer Perceptrons

CLOUD FREE MOSAIC IMAGES

A Time Series ANN Approach for Weather Forecasting

Use of Artificial Neural Network in Data Mining For Weather Forecasting

Lecture 6. Artificial Neural Networks

Stock Market Prediction System with Modular Neural Networks

STATIC SIGNATURE RECOGNITION SYSTEM FOR USER AUTHENTICATION BASED TWO LEVEL COG, HOUGH TRANSFORM AND NEURAL NETWORK

Supply Chain Forecasting Model Using Computational Intelligence Techniques

Neural Networks: a replacement for Gaussian Processes?

DYNAMIC LOAD BALANCING OF FINE-GRAIN SERVICES USING PREDICTION BASED ON SERVICE INPUT JAN MIKSATKO. B.S., Charles University, 2003 A THESIS

REVIEW OF HEART DISEASE PREDICTION SYSTEM USING DATA MINING AND HYBRID INTELLIGENT TECHNIQUES

Recurrent Neural Networks

Design call center management system of e-commerce based on BP neural network and multifractal

Prediction of Cancer Count through Artificial Neural Networks Using Incidence and Mortality Cancer Statistics Dataset for Cancer Control Organizations

Back Propagation Neural Network for Wireless Networking

Artificial Neural Network and Non-Linear Regression: A Comparative Study

4.1 Learning algorithms for neural networks

A Multi-level Artificial Neural Network for Residential and Commercial Energy Demand Forecast: Iran Case Study

Time Series Data Mining in Rainfall Forecasting Using Artificial Neural Network

SURVIVABILITY ANALYSIS OF PEDIATRIC LEUKAEMIC PATIENTS USING NEURAL NETWORK APPROACH

Introduction to Machine Learning Using Python. Vikram Kamath

Machine learning in financial forecasting. Haindrich Henrietta Vezér Evelin

Spatio-Temporal Neural Data Mining Architecture in Learning Robots

Numerical Algorithms Group

Method of Combining the Degrees of Similarity in Handwritten Signature Authentication Using Neural Networks

W6.B.1. FAQs CS535 BIG DATA W6.B If the distance of the point is additionally less than the tight distance T 2, remove it from the original set

Linear Threshold Units

STA 4273H: Statistical Machine Learning

A Simple Feature Extraction Technique of a Pattern By Hopfield Network

Iranian J Env Health Sci Eng, 2004, Vol.1, No.2, pp Application of Intelligent System for Water Treatment Plant Operation.

Monotonicity Hints. Abstract

Quality Assessment in Spatial Clustering of Data Mining

Neural Networks for Machine Learning. Lecture 13a The ups and downs of backpropagation

Chapter 2 The Research on Fault Diagnosis of Building Electrical System Based on RBF Neural Network

International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014

Impelling Heart Attack Prediction System using Data Mining and Artificial Neural Network

PLAANN as a Classification Tool for Customer Intelligence in Banking

TRAIN AND ANALYZE NEURAL NETWORKS TO FIT YOUR DATA

A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images

Pattern Recognition Applications in Securities Trading

Transcription:

SUCCESSFUL PREDICTION OF HORSE RACING RESULTS USING A NEURAL NETWORK N M Allinson and D Merritt 1 Introduction This contribution has two main sections. The first discusses some aspects of multilayer perceptrons, while the second outlines an application - namely the prediction of horse racing results. 2 Multi-layer perceptrons Most application work within neural computing continues to employ multi-layer perceptrons ( MU). Though many variations of the fully interconnected feed-forward MLP, and even more variations of the back propagation learning rule, exist; the first section of this paper attempts to highlight several properties of these standard networks. 2.1 Adaptive Mappers t extrapolation, training samples test samples Fig. 1 Fig. 2 MLPs act as adaptive mappers (Fig. 1) that learn, through supplied examples, to map from one function in the space to another in the output space. This ability can be exploited for supervised pattern classification. Note that the learning must be supervised, in that pattern vectors and output target vectors need to be presented in pairs. Such supervised learning implies that we already possess an a priori model of the underlying processes involved. Fig. 2 illustrates this mechanism for the onedimensional situation. The MLP interpolates a smooth function between the supplied training examples. Note that the Nyquist's sampling theory has a part to play, in that training samples must be sufficiently close to each other and well distributed. Therefore the MLP acts as an interpolator and, possibly, a poor extrapolator. The use of sigmoid compressive functions in the processing elements or neurons of the network help to ensure a smooth interpolation curve. For many applications this is ideal. Most physical phenomena are convoluted by Gaussian shaped corrupting functions. But, Department of Electronics, University of York, York, YO1 5DD

the application of sigmoid compressive functions to training data, which we may term logical (see Fig. 3), is much less likely to be successful. lopology exdu8ive 4 R rnnhod CIWIH. A tnining snnples test samples 2.2 Network complexity Fig. 3 Fig 4 The abilities of MLPs in isolating regions of pattern space as a function of the number of layers are well known. Fig. 4 illustrates the general form of the discriminant functions possible. These are for the case of a hard non-linearity function in the neurons where, in the n-dimensional situation, hyper-planes can be constructed. For a single layer network, only a simple partitioning of the space is possible; for two layers, convex open or closed region of the space; and for three layers, any arbitrary shaped region or regions. The ability to cope with hidden structures (ie, the OR-ing of disjoint clusters) is one of the most significant abilities of h4lps. This is only possible in a supervised learning system. The use of smooth non-linearities that permit the application of the generalised delta rule (ie, differentiable) cause the region boundaries to become fuzzy (Fig. 51, which has implications for the generalisation abilities of the resultant MLP. For example, a test pattern can contribute to two or more regions. Extensions to a greater number of layers may help in successful learning since we are expecting a reduced complexity of non-linear mapping in each layer. A question often discussed concerns the number of hidden units (ie, the number of neurons in layers which are not or output layers). In general, there is no way of predicting this number, as it depends on the detailed shape of the pattern clusters in the appropriate space. The number of units is not, as is often suggested, related to the inherent dimensionality of the pattern vectors. The situation is illustrated in Fig. 6 for two cases of similar two-dimensional pattern clusters. It should be noted that the hyperplane discriminant boundaries extend throughout the pattern space. This may be a contributory factor to the difficulties in scaling MLPs to more complex problems. If there are no means to deade the number of hidden units, what effect does changing the number of units have? Too few and the system will be unable to classify

-t hard non-linesnty Fig. 5 Fig 6 speafic clusters (Fig. 7a), that is it will over generalise; too many and the system will be unable to generalise at all (Fig. 7c). Only you, as the designer, will be able to recognise satisfactory performance (Fig. 7b). (a) (b) 2.3 Learning Fig 7 Learning is slow; usually many thousand cycles of training are required. It is difficult to move correctly discriminant functions (ie, hyperplanes) in space. In pattern clustering techniques, both traditional classification and neural network methods, single points are moved in space. This is much easier and, in part, explains the somewhat short learning times of such methods. The error surface (ie, the difference between the current output and target vectors, for each layer) can be a complex high dimensional function. As learning is essentially a method of gradient descent optimisation, then there is a high probability of becoming captured in a local minimum of this function. Varying the adaptation step size, use of momentum terms or higher order derivatives or use of added noise can help in overcoming this difficulty. A typical cross-section of the error surface is usually portrayed as in Fig. 8a; however it is more likely to take on the form of Fig. 8b (especially for the logicnl data as mentioned in 2.1). It is easy to choose a direction to go in order to find sea-level if you are stood on the top of a mountain. Much more difficult if you are stood on a horizontal plain. 4/ 3

2? - f D iterations (b) Fig 8 Fig 9 There is much evidence to suggest that MLPs learning algorithms are chaotic; certainly they are ill-conditioned. The same training schedule appiied to the same network using the same data can lead to very different results. Not all attempts at learning results in the graceful decrease in error function with training time (Fig. 9). There are many reasons for unsuccessful learning - incorrect learning parameters, incorrect network complexity, over-learning (ie, over specialisation), inconsistent training data. It is difficult to diagnosis the reasons for failure, but experience has often highlighted problems in the training and application data sets. Variations on training include preprocessing the data, prior to applying it to the network. There are considerable advantages in transforming the raw data into a set of basis functions, as we now require less complex non-linear mapping from the MLP. 2.4 Concluding remarks The previous sections may seem over critical of MLPs but they are capable of providing a very useful pattern classification scheme for many situations - out performing traditional techniques. However, they are not a panacea for all problems. Perhaps their performance can be summarized as: MLPs are good but not that good 4/4