Pattern recognition using multilayer neural-genetic algorithm



Similar documents
A Robust Method for Solving Transcendental Equations

An intelligent stock trading decision support system through integration of genetic algorithm based fuzzy neural network and articial neural network

Lecture 6. Artificial Neural Networks

Analecta Vol. 8, No. 2 ISSN

Hybrid Evolution of Heterogeneous Neural Networks

ISSN: ISO 9001:2008 Certified International Journal of Engineering Science and Innovative Technology (IJESIT) Volume 2, Issue 3, May 2013

Introduction to Machine Learning and Data Mining. Prof. Dr. Igor Trajkovski

PLAANN as a Classification Tool for Customer Intelligence in Banking

SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS

Keywords: Image complexity, PSNR, Levenberg-Marquardt, Multi-layer neural network.

Numerical Research on Distributed Genetic Algorithm with Redundant

Chapter 4: Artificial Neural Networks

An Introduction to Neural Networks

Cash Forecasting: An Application of Artificial Neural Networks in Finance

ARTIFICIAL NEURAL NETWORKS FOR ADAPTIVE MANAGEMENT TRAFFIC LIGHT OBJECTS AT THE INTERSECTION

OPTIMUM LEARNING RATE FOR CLASSIFICATION PROBLEM

A Simple Feature Extraction Technique of a Pattern By Hopfield Network

FRAUD DETECTION IN ELECTRIC POWER DISTRIBUTION NETWORKS USING AN ANN-BASED KNOWLEDGE-DISCOVERY PROCESS

Prediction of Software Development Modication Eort Enhanced by a Genetic Algorithm

International Journal of Software and Web Sciences (IJSWS)

Architecture bits. (Chromosome) (Evolved chromosome) Downloading. Downloading PLD. GA operation Architecture bits

APPLICATION OF ARTIFICIAL NEURAL NETWORKS USING HIJRI LUNAR TRANSACTION AS EXTRACTED VARIABLES TO PREDICT STOCK TREND DIRECTION

Optimum Design of Worm Gears with Multiple Computer Aided Techniques

Feedforward Neural Networks and Backpropagation

A Parallel Processor for Distributed Genetic Algorithm with Redundant Binary Number

Power Prediction Analysis using Artificial Neural Network in MS Excel

EFFICIENT DATA PRE-PROCESSING FOR DATA MINING

Vol. 35, No. 3, Sept 30,2000 ملخص تعتبر الخوارزمات الجينية واحدة من أفضل طرق البحث من ناحية األداء. فبالرغم من أن استخدام هذه الطريقة ال يعطي الحل

Chapter 2 The Research on Fault Diagnosis of Building Electrical System Based on RBF Neural Network

American International Journal of Research in Science, Technology, Engineering & Mathematics

Supply Chain Forecasting Model Using Computational Intelligence Techniques

Application of Neural Network in User Authentication for Smart Home System

A Fast Computational Genetic Algorithm for Economic Load Dispatch

A RAPID METHOD FOR WATER TARGET EXTRACTION IN SAR IMAGE

Multiple Layer Perceptron Training Using Genetic Algorithms

A Content based Spam Filtering Using Optical Back Propagation Technique

MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL

D A T A M I N I N G C L A S S I F I C A T I O N

Price Prediction of Share Market using Artificial Neural Network (ANN)

Performance Evaluation of Artificial Neural. Networks for Spatial Data Analysis

Neural Network and Genetic Algorithm Based Trading Systems. Donn S. Fishbein, MD, PhD Neuroquant.com

6.2.8 Neural networks for data mining

Machine Learning and Data Mining -

Forecasting Trade Direction and Size of Future Contracts Using Deep Belief Network

Comparison of K-means and Backpropagation Data Mining Algorithms

Correspondence should be addressed to Chandra Shekhar Yadav;

Role of Neural network in data mining

1. Introduction MINING AND TRACKING EVOLVING WEB USER TRENDS FROM LARGE WEB SERVER LOGS. Basheer Hawwash and Olfa Nasraoui

Building an Advanced Invariant Real-Time Human Tracking System

Face Recognition For Remote Database Backup System

Keywords revenue management, yield management, genetic algorithm, airline reservation

Lab 4: 26 th March Exercise 1: Evolutionary algorithms

Neural network software tool development: exploring programming language options

Neural Network Design in Cloud Computing

Impact of Feature Selection on the Performance of Wireless Intrusion Detection Systems

Data quality in Accounting Information Systems

SUCCESSFUL PREDICTION OF HORSE RACING RESULTS USING A NEURAL NETWORK

Memory Allocation Technique for Segregated Free List Based on Genetic Algorithm

Artificial Neural Network for Speech Recognition

A hybrid Approach of Genetic Algorithm and Particle Swarm Technique to Software Test Case Generation

A Prediction Model for Taiwan Tourism Industry Stock Index

A Genetic Algorithm Approach for Solving a Flexible Job Shop Scheduling Problem

Introduction To Genetic Algorithms

Manufacturing and Fractional Cell Formation Using the Modified Binary Digit Grouping Algorithm

Simplified Machine Learning for CUDA. Umar

Comparison of Major Domination Schemes for Diploid Binary Genetic Algorithms in Dynamic Environments

COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS

College of information technology Department of software

Maintenance Scheduling Optimization for 30kt Heavy Haul Combined Train in Daqin Railway

Neural Network Predictor for Fraud Detection: A Study Case for the Federal Patrimony Department

NEURAL NETWORKS IN DATA MINING

A Multi-level Artificial Neural Network for Residential and Commercial Energy Demand Forecast: Iran Case Study

Performance Evaluation On Human Resource Management Of China S Commercial Banks Based On Improved Bp Neural Networks

A Genetic Algorithm-Evolved 3D Point Cloud Descriptor

degrees of freedom and are able to adapt to the task they are supposed to do [Gupta].

Artificial Intelligence and Machine Learning Models

A Concise Neural Network Model for Estimating Software Effort

Evaluation of Feature Selection Methods for Predictive Modeling Using Neural Networks in Credits Scoring

K-Means Clustering Tutorial

Stock Market Index Prediction by Hybrid Neuro- Genetic Data Mining Technique

Open Access Research on Application of Neural Network in Computer Network Security Evaluation. Shujuan Jin *

Using quantum computing to realize the Fourier Transform in computer vision applications

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals

Adaptive Radio Resource Management for GSM/GPRS Networks

A New Approach For Estimating Software Effort Using RBFN Network

Automatic Inventory Control: A Neural Network Approach. Nicholas Hall

Time Series Data Mining in Rainfall Forecasting Using Artificial Neural Network

Neural network models: Foundations and applications to an audit decision problem

Data Mining using Artificial Neural Network Rules

Goldberg, D. E. (1989). Genetic algorithms in search, optimization, and machine learning. Reading, MA:

Copyright 2005 IEEE. Reprinted from IEEE MTT-S International Microwave Symposium 2005

DIAGONAL BASED FEATURE EXTRACTION FOR HANDWRITTEN ALPHABETS RECOGNITION SYSTEM USING NEURAL NETWORK

Retrospective Test for Loss Reserving Methods - Evidence from Auto Insurers

Predicting the Risk of Heart Attacks using Neural Network and Decision Tree

Multiobjective Multicast Routing Algorithm

Transcription:

Neurocomputing 51 (2003) 237 247 www.elsevier.com/locate/neucom Pattern recognition using multilayer neural-genetic algorithm Yas Abbas Alsultanny, Musbah M. Aqel Computer Science Department, College o Computer and Inormation Technology, Applied Science University, P.O. Box17, Amman 11931, Jordan Received 2 July 2001; accepted 8 May 2002 Abstract The genetic algorithm implemented with neural network to determine automatically the suitable network architecture and the set o parameters rom a restricted region o space. The multilayer neural-genetic algorithm was applied in image processing or pattern recognition, and to determine the object orientation. The library to cover the views o object was build rom real images o (10 10) pixels. Which is the smallest image size can be used in this algorithm to recognize the type o aircrat with its direction The multilayer perecptron neural network integrated with the genetic algorithm, the result showed good optimization, by reducing the number o hidden nodes required to train the neural network (the number o epoch s reduced to less than 50%). One o the important results o the implemented algorithm is the reduction in the time required to train the neural network. c 2002 Elsevier Science B.V. All rights reserved. Keywords: Neural network; Genetic algorithm; Image processing; Pattern recognition 1. Introduction Computer imaging separated into two elds: Computer vision. Image processing. Computer vision is computer imaging where the application does not involve a human being in the visual loop. Image analysis involves the examination o the image Corresponding author. Fax: 5232899. E-mail address: alsultanny@hotmail.com (Y.A. Alsultanny). 0925-2312/03/$ - see ront matter c 2002 Elsevier Science B.V. All rights reserved. PII: S0925-2312(02)00619-7

238 Y.A. Alsultanny, M.M. Aqel / Neurocomputing 51 (2003) 237 247 Input Image Pattern Preprocessing Classiication Image Pattern (Samples) Supervised or Unsupervised Learning Output Decision Fig. 1. Block diagram o decision-theoretical pattern recognition system. data to acilitate solving a vision problem. The image analysis process involves two other topics: eature extraction is the process o acquiring higher-level image inormation, such as shape or color inormation, and pattern classication is the act o taking this higher-level inormation identiying objects within image. The block diagram o decision-theoretical pattern recognition shown in Fig. 1 represented by three primary stages [11,6]: Preprocessing Learning Classication Preprocessing is used to remove noise and eliminate irrelevant, visually unnecessary inormation. Learning stage involves either supervised or unsupervised learning. Classication stage classied the pattern, which is preprocessed in the rst stage, by computing the output and considering the maximum output is the more similar image in database library to the unknown input pattern image. 2. Multilayer perecptron neural network (MLP) The three layers network with sigmoid activation unction is capable to produce a decision unction with enough hidden units is shown in Fig. 2. The input layer contains the input nodes interacts with the outside environment. The input units are only buer units that pass the signal without changing it. The hidden layer size is let to the appreciation o the user that will estimate the number o hidden nodes by his experience or by trial and error. This number must not be too large to avoid waste o computations and slow convergence process and not too low to make the network capable to absorb the set o training pixels. The output layer represent the number o nodes, which is equal to the number o classes, each output node representing a class [12,7,4,8]. 3. Genetic algorithm(ga) The parameters which needs to be searched by GA in the training are the range o initial weights (R), initial values o learning rate and momentum rate, and the network architecture. The initial weights are generated using a uniorm random number generator

Y.A. Alsultanny, M.M. Aqel / Neurocomputing 51 (2003) 237 247 239 x 1 E 1 y 1 t 1 - x 2 E 2 y 2 t 2 - x N y M - t M E M Input layer hidden layer output layer error target Fig. 2. Multilayer neural network architecture x 1 ;x 2 ;:::;x N = input nodes, y 1 ;y 2 ;:::;y M = output nodes, E 1 ;E 2 ;:::;E M = error, t 1 ;t 2 ;:::;t M = target. Selection Population (Chromosomes) Phenotype Replacement Mating pool (Parent) Fitness Objective Function Genetic Operation Fitness Sub-population (Ospring) Phenotype Fig. 3. Cycle o GA. in the range [ R; R]. The network architectures described by the number o hidden layer (h), the number o nodes in each layer (n i ;i=0; 1;:::;h). The GA proposed as searching process based on the laws o the natural selection and genetics. This algorithm which is used rstly to obtain the optimal values o the hidden nodes and the initial values o learning rate (R) and momentum top-level description o a simple GA is shown below [5,3]: 1. Randomly generate an initial population P 0 =(a 0 1 ;a0 2 :::a0 ).

240 Y.A. Alsultanny, M.M. Aqel / Neurocomputing 51 (2003) 237 247 Start First stage Genetic Algorithm Second stage MLP algorithm Initialize a population o chromosomes as current generation Initialize weight using the chromosome with the smallest itness unction Calculate the itness (objective) unction or each chromosome Feed orward step o MLP neural net using sigmoid activation unction Apply the crossover and mutation operators to the parent chromosomes and produced a new generation Calculate the total sum squared error between actual and desired output Replace the new generation as current generation Update the weight using back-propagation algorithm No Stop the irst learning End Stop No Yes Yes The control low o the irst learning stage The control low o the second learning stage Fig. 4. The neural-genetic learning algorithm. 2. Compute the tness (a t i ) o each chromosome at i in the current population P t. 3. Create new chromosome P t o mating current chromosomes, applying mutation and recombination as the parent chromosomes mate. 4. Delete numbers o the population to make room or new chromosomes. 5. Compute the tness o (a t i ), and insert these into population. 6. Increment number o generation, i not (end-test) go to step 3, or else stop return the best chromosome. The GA cycle is shown in Fig. 3. From which the phenotype represents the coding which is used to represent the chromosomes. The mating pool represents the summing o chromosomes that represents the solution o the problem. Ospring represents new chromosomes result by mutation and crossover operation. Lastly, the objective unction represents the unction that measured the perormance o the origin unction [10,9,1].

Y.A. Alsultanny, M.M. Aqel / Neurocomputing 51 (2003) 237 247 241 4. Neural-genetic network learning algorithm The proposed learning algorithm using genetic algorithm with multilayer neural networks or pattern recognition is presented in Fig. 4. It consists o two learning stages. The rst learning using genetic algorithm with the eed-orward step uses stage to accelerate the whole learning process. The genetic algorithm perorms global search and seeks near-optimal initial point (weight vector) or the second stage. Where, each chromosome is used to encode the weights o neural network. The tness (objective) unction or genetic algorithm is dened as the total sum squared system error (TSSE) o the corresponding neural network. Thereore, it becomes an unconstrained optimization problem to nd a set o decision variables by minimizing the objective unction. The best tness unction value in a population is dened as smallest value o the objective in the current population. Let us consider the neural-genetic algorithm as ollows: =* First learning stage *= Initialize the chromosomes randomly as current generation, initialize working parameters, and set the rst chromosome as the best chromosome. a-for i = 0 to population size, do concurrently: -Initialize sub total tness to zero and sub best chromosome as null. b-for j = 1 to Length o chromosomes, do sequentially: -Perorm eed-orward procedure o the MLP (using sigmoid activation unction). -Calculate the objective unction (system error or the neural network). Next j -Calculate the total tness by accumulating sub total tness. c-store the best chromsome as sub best chromosome. Next i Guarded section 1-entry d-compare the sub best chromsome with each other and set the best sub best chromosome as the best chromosome. Guarded section 1-end Do e-for i = 0 to (population size=2), do concurrently: -Initialize sub total tness to zero and sub best chromosome as null. -For j = 1 to length o chromosome, do sequentially: Select parent using roulette wheel parent selection. Apply two-point, multipoint, or uniorm crossover and mutation to the parents.

242 Y.A. Alsultanny, M.M. Aqel / Neurocomputing 51 (2003) 237 247 -For k = 1 to length o chromosome, do sequentially: Perorm eed-orward procedure o the MLP to parent chromosomes. Calculate the objective unction (system error or the neural network) or parent chromosomes. Next k -Calculate the sub total tness by accumulating objective o each chromosome. -Store the best chromosome as sub best chromosome. Next j g-replace the old generation by the new generation WHILE (stopping criteria). =* Second learning stage *= -Set the best chromosome as the initial weight vector, set up the topologic structure o neural network. -Compute the actual output o the eed-orward MLP neural network. -Compute the error between desired output and actual. -Updates the weight using back-propagation algorithm. END 5. Implementation and results Multilayer neural networks are trained to (12) images o sizes 10 10 pixels, which represents the smallest image size required or the aircrat recognition, where we can recognize the aircrat with their direction. The initial values such as the number o hidden nodes, learning techniques, and input coding in addition to initialization weight have been dened. To test the ability o net to pattern recognition in recall or testing phase, the ailure rate proposed as the measurement to the sensitivity o neural network is dened as Failure rate (FR) Total No: o tested images No: o classied images = 100%: Total number o test image Fig. 5 shows the images o 256 gray levels used to test the proposed algorithm. Test 1: MLP network Training parameters: Image size = 10 10 pixels Number o images = 12 Desired Accuracy = 90% Hidden nodes = 60

Y.A. Alsultanny, M.M. Aqel / Neurocomputing 51 (2003) 237 247 243 Fig. 5. Groups o images o 256 gray level used in testing o the neural-genetic algorithm. (a) Four images to our directions o the same groups o aircrat; (b) Four images to our directions o the same aircrat with missile; (c) Four images to our directions o the same aircrat. Learning rate = 0:6 Momentum = 0:06 Training result are: Epoch s 75 150 225 300 366 Failure rate 83.33% 58.33% 50% 50% 0.00% TSSE 0.7252 0.6747 0.3647 0.3613 5:7 10 5 Test 2: Neural-genetic network learning algorithm The neural-genetic network learning algorithm applied to the same images. A threelayer neural network with one hidden layer was used to learn this problem. The total number o links in these three-layer neural networks is 6600.

244 Y.A. Alsultanny, M.M. Aqel / Neurocomputing 51 (2003) 237 247 Training parameters: Population size = 12 Mutation rate = 0:05 Crossover rate = 0:46 Real coding to the chromosomes Length o chromosomes = 6600 Number o generation = 8 Desired accuracy = 90% Training results are: No. o generation 2 4 6 8 Total Fitness 6.5653 6.4254 5.9028 5.6283 Training results are: Epoch s 25 50 75 100 125 158 Failure rate 83.33% 75% 66.67% 58.33% 50% 0.00% TSSE 0.9831 0.8616 0.5054 0.3651 0.216 0.0008 The system error on the rst test by using the MLP learning neural network algorithm was 5:73 10 5 ater 366 iteration. Ater 8 generations, the rst learning stage o neural-genetic network learning algorithm met stopping criteria. The best average tness unction has value o about 5.6283. The result o the rst learning stages is used as an initial weight vector or the second learning stage. Ater 158 iterations, the system error in the neural-genetic neural network learning algorithm converges to a value o less than 0.0008 and the algorithm achieves 100% to train all the images. The perormance o hybrid neural-genetic networks and MLP neural net are shown in Fig. 6. 6. Conclusions The neural network depends on the training conditions (hidden node number, learning rate, momentum, and initial weight). The minimum size used to recognize the type o aircrat and its direction is (10 10) pixels. This technique can be used to detect the stealth aircrat, which cannot be detected by the radar signal. The learning time depends strongly on the resolution o image, hidden nodes and training pixels. The proposed genetic algorithm presented to nd suitable network, which was able to perorm well within the specied tolerance e.g. zero ailure rate or MLP net. As

Y.A. Alsultanny, M.M. Aqel / Neurocomputing 51 (2003) 237 247 245 TSSE & FR[%] (a) 100 80 60 40 20 0 TSSE[%] FR[%] 0 100 200 300 400 Epoch's TSSE & FR[%] (b) 100 90 80 70 60 50 40 30 20 10 0 0 50 100 150 200 Epoch's TSSE[%] FR[%] Fig. 6. (a) Perormance o MLP neural net. (b) Perormance o neural-genetic learning algorithm. ar as training procedures are concerned, we have identied some important learning phenomenon e.g. the eect caused by learning rate, number o hidden units, training o under units, training sets, etc. Thus, the ratio o reduction on the number o epoch s is least or equal to 50% i we used the genetic algorithm. Integrating the genetic algorithm with MLP network, the ollowing observations are the conclusions can be drawn: The results o neural network learning are sensitive to the initial value o the weight vector. A genetic algorithm is employed to perorm global search and to seek a good starting weight vector or subsequent neural network learning algorithm. The result is an improvement in the convergence speed o the algorithm. The problem o entrapment in a local minimum is encountered in gradient-based neural network learning algorithm. In the neural-genetic learning algorithm presented, this problem is circumvented using a genetic algorithm, which is guided by the tness unction o population rather than gradient direction. Ater several iterations o the global search, the rst learning stage returns a near-global optimum point that is used as the initial weight vector or the second learning stage.

246 Y.A. Alsultanny, M.M. Aqel / Neurocomputing 51 (2003) 237 247 7. For urther reading The ollowing reerence may also be o interest to the reader: [2]. Acknowledgements The authors are very grateul to the useul comments given by anonymous reviewers, which helped greatly in revising this work. Reerences [1] D. Anthony, E. Hines, The use o genetic algorithms to learn the most appropriate inputs to neural network, Application o the International Association o Science and Technology or Development-IASTED, June, 1990, 223 226. [2] T.N. Bui, B.R. Moon, Genetic algorithms and graph partitioning, IEEE Trans. comput. 45 (7) (1996) 841 855. [3] L. Davis, Hand book o Genetic Algorithms, Van Nostrand Reinhold, New York, 1991. [4] L. Fauselt, Fundamentals o Neural Networks, Prentice-Hall, International Inc., Englewood Clis, NJ, 1994. [5] D.E. Goldberg, Genetic Algorithms in Search, Optimization, and Machine Learning, Addison-Wesley Pub. Comp. Inc., Reading, MA, 1989. [6] R. Gonzalez, P. Wintz, Digital Image Processing, Addison-Wesley Pub. Comp. Inc., Reading, MA, 1992. [7] R.L. Harvey, Neural Networks Principles, Prentice-Hall, Englewood Clis, NJ, 1994. [8] R.P. Lippmann, Introduction to computing with neural nets, IEEE Trans. Acoust. Speech Signal Process. 4 (87) (1987) 3 22. [9] M. Mitchill, An introduction to genetic algorithms, Massachusetts Institute o Technology, England, 1996. [10] S.c. Ng, S.H. Leung, C.Y. Chung, A. Luk, W.H. Lau, The genetic search approach, IEEE Signal process. Mag. 13 (6) (1996) 38 46. [11] S.E. Umbaugh, Computer Vision and Image Processing A practical approach using CVIP tools, Prentice-Hall PTR, Englewood Clis, NJ, 1998. [12] J.M. Zurada, Introduction to Articial Neural Systems, Jaico Pub. House, Bombay, India, 1997. Yas Abbas K. Alsultanny received the B.S. degree rom the Technology University, Control and Systems Engineering Department, Iraq in 1982, the M.S. degree, in Computer Engineering rom the Technology University, Iraq 1991, and the Ph.D. degree, in image processing rom Baghdad University, Iraq, in 1997. From 1997 to 2000, he was an assistant proessor at the Computer Engineering Department in Baghdad, and a member in the scientic research council in Baghdad. He supervised 13, M.S. thesis. From 2000 he was an assistant proessor at the computer Science department, in the Computer Science and Inormation Technology Department in the Applied Science University, in Jordan, he has been a plenary speaker at various conerences. His research interested, include, image processing, pattern recognition, uzzy set, neural network, wavelet, and GIS.

Y.A. Alsultanny, M.M. Aqel / Neurocomputing 51 (2003) 237 247 247 Musbah M. Aqel was born in Hebron=West Bank, in 1959, He received the B.E. degree in electronics & Communication rom regional engineering college, University o Kashmir, India, in 1987, and the M.Tech. and Ph.D. Degrees in Computer Science rom School o Computer and Systems Sciences, Jawaharal Nehru University, New Delhi, India, in 1989 and 1993, respectively. From 1994 to 1999 he was an Assistant Proessor at the Department o Computer Science and Inormation Systems, Faculty o Basic Science at Applied Science University, Amman, Jordan, From 1999 to 2001 was a chairman o Computer Science and Inormation systems Department. He is a member o organization steering, and technical Committees o many national and international conerences, the advisory Board o Jordan Journal o Applied Science, Jordan Engineering Association, and Jordan Computer Society. He is Currently Dean o the Faculty o Computer Science and Inormation Technology and Associate Proessor with the Department o Computer Science at Applied Science University, Amman, Jordan. His research interests include Algorithms Design and Analysis, Database Systems, Simulation and Modeling, Pattern Recognition, and Image Processing.