Neural Networks for Pen Characters Recognition
|
|
|
- Grace McCoy
- 9 years ago
- Views:
Transcription
1 Neural Networks for Pen Characters Recognition Carlos Agell Department of Electrical Engineering and Computer Science University of California, Irvine Irvine, CA Abstract This paper presents experiments that compare different kinds of features for isolated handwritten character recognition in a Neural Network environtment. The selection of valuable features is crucial in character recognition, therefore a new and meaningful set of features, the Uniform Differential Normalized Coordinates (UDNC), is introduced. These features are shown to improve the recognition rate using simple classification algorithms so they are used to train a Neural Network and test its performance on a reduced database. 1 Introduction Character recognition, both human and machine generated, is a wide and largely studied field in Machine Learning. In fact, nowadays many commercial scanners use Optical Character Recognition (OCR) systems that output a character string having an image of typed text as input. In the age of Personal Digital Assistants (PDA) and tablet PCs, there is an upgrowing need and interest for handwritten character recognition. This kind of recognition, slightly different from the general OCR approach, will be the scope of this paper. This paper proposes to solve the classification of isolated handwritten characters and digits of the UJI Pen Characters Data Set[1] using Neural Networks. The data [6] consists of samples of 26 characters and 10 digits written by 11 writers on a tablet PC. The characters (in standard UNIPEN format 1 ) are written both in upper and lower case and there is a whole two set of characters per writer. So the output should be in one of the 35 classes. The ultimate objective is building a writer independent model for each character. The approach of this paper leans in two main points: extracting meaningful features and training a neural network to correctly classify the data. The rest of the paper is organized as follows: after this introduction there is a brief review of the state of the art in character recognition and the different approaches taken to solve this problem. Section 3 is dedicated to select the appropriate features, where a new set 1 See the UNIPEN Project history in
2 of features for standard UNIPEN characters is introduced. Following there is a section describing the implementation the approach in this paper. Section 5 shows the analysis of the results from the implementation. Finally the conclusion points out the main novelties reported in this paper. 2 Previous Work Several approaches might be found in literature to solve the character recognition problem using different Machine Learning techniques and algorithms. This paper will only discuss about isolated character recognition. On the one hand the typical character recognition has been attacked with tools such as simple Nearest Neighbour[4], Support Vector Machines (SVM)[8] and Neural Networks[3, 4], that leaded to impressive results on different databases. But this approaches are all used when the input are images, such as the classical MNIST database of mail zip codes [5]. On the other hand, for the specific case of UNIPEN character recognition, there are diverse approaches feature-wise and algorithm-wise. Regional-Fuzzy Representation (RFR) lead to impressive recognition rates when using them as inputs for a Neural Network [7]. The experiment framework for this case, though, worked only with lower-case, upper-case and digits separately. A whole discussion about feature selection can be found in [2], where part of the work of this project was inspired on. 3 Feature Selection A correct feature selection is crucial in any pattern recognition problem and handwritten character recognition is not an exception. If we want to classify we need to measure in dimensional space common for all the characters (recall that the number of coordinates in a UNIPEN representation varies depending on the character). This section describes the two main approaches proposed as features for the UJI database: making images from the features and processing the UNIPEN data to make meaningful features. We also introduce the Uniform Differential Normalized Coordinates (UDNC) as part of the second approach. Finally, we compare all the features using a simple classification algorithm. 3.1 Making Images from the Features Making images out of the UNIPEN coordinates seems a correct approach for character recognition as there is a strong background in this problem using those kind of features. In order to get feature images to use for recognition we first center the data and linearly interpolate on a grid so that we can reconstruct the original character. Computationallywise we would like to have the fewer dimensions as possible so that the classification process does not take long, that is why we could think about decimation. But decimation makes no sense in a linearly interpolated image because the connections between points will be lost. For this reason we think it is useful to use mathematical morphology, to erode the initial image and then decimate it. This way we can euristically lower the dimensions of the image, up to a representation of pixel image which seems reasonally good.
3 Figure 1: Coordinates of the UNIPEN format data (left), linear interpolation (center), and final pixel image Figure 2: Coordinates of the UNIPEN format data (left) and uniform sampling using 35,100 and 200(right) The initial coordinates, the linear interpolation and the final pixel image are depicted in figure 1. However there are several drawbacks in the use of these kind of features. First of all they are not actually for online character recognition as they need to be converted to images for further classification. This means postprocessing and not online processing as it is pointed out in the goal of the UJI database [6]. And secondly we are missing an important piece of information when converting to images: there is a time information hidden (some coordinates are generated first than others). 3.2 UNIPEN Feature Processing With the improvement of the features in the previous section as a goal, this section offers different approaches to the features to use for the UNIPEN handwritten character recognition problem. Some ideas are introduced to face the problems UNIPEN format imposes. The section culminates with the definition of the Uniform Differential Normalized Coordinates which will overcome all the UNIPEN format problems UNIPEN format drawbacks The main drawback of the UNIPEN format is that is not a single dimensional format for all the data. Each datapoint has different number of coordinates. The previous section faced this problem by expanding the number of dimensions but somehow, this approach introduces redundant and non useful data. We propose a uniform sampling for the UNIPEN format. This means to get the same number of coordinates for all the characters by resampling them using linear interpolation. Figure 2 shows this idea.
4 Once we have the dimensionality problem solved, we need to overcome another one, position and rotation (maybe the character is slighly in italics). This fact is also present in the image features, though is slightly overcome by centering. The problem could be solved using relative coordinates instead of absolute ones. Finally there is another drawback, the length of each character is different. This means that when if we uniformly resample we are not taking into account the total lenght of the character (sum of the distances between all the points). Normalization could address this fact. The three improvements cited are putted together in the Uniform Differential Normalized Coordinates Uniform Differential Normalized Coordinates The Uniform Differential Normalized Coordinates (UDNC) are defined as the difference between the uniformly resampled coordinates normalized by the length of the character. The construction of this kind of coordinates follows the following algorithm, assuming we have the original UNIPEN format data. Uniformly resample each character using linear interpolation (U i, U R n+2 2 ). Calculate the difference between datapoints (UD i = U i+1 U i, D R n 2 ). Normalize by the total length of the character, UDN i = R n 2 UD i i UDi, UDN It must be pointed out that some of the characters have multiple traces (to write an X you need to put the pen up from the paper). Those characters are uniformly resampled proportionally to the length of each of the traces. The parameter n can be choosen, so the final dimensions will be 2n. For this work and the particular case of the characters the value n = 35 allowed a correct reconstruction of all the characters. These UDC coordinates are normalized, invariant to rotations and displacements, and have the same dimensions to represent all the data. Furthermore they can be applied to recognize handwritten characters online. 3.3 Feature Comparison As training a Neural Network can be computationally expensive, this section exposes the comparison of different features using simple classification algorithms. These simple classification algorithms are computationally reasonable and allow crossvalidation to see which features performs better. The best behaving features will further be used by the Neural Network implementation to get a better classification. Table 1 shows the performance of the features tried in this paper for leave-one-out crossvalidation. The simple algorithms applied are mainly Nearest Neighbor [4], Nearest Mean and its SVD equivalent. Nearest Mean is nothing more than choosing the closest mean to the input data, when the mean has been calculated using all the data except for the characters of a particular writer. The SVD equivalent is nothing more than using the projections of the data on the first k eigenvectors and using them as features. This last algorithm used k = 100 and was an approach to reduce dimensionality.
5 Table 1: Recognition Rates for different algorithms and features Features Algorithm Max Mean Image Nearest Neighbour 45.97% 31.82% Image Nearest Mean 62.19% 43.84% Image SVD 1-Nearest Mean 53.23% 40.32% Uniform Coordinates(n = 35) 1-Nearest Neighbour 77.42% 63.56% Differential Angles(n = 35) 1-Nearest Neighbour 69.35% 55.65% Uniform Differential(n = 35) 1-Nearest Neighbour 83.06% 67.08% UDN Coordinates(n = 35) 1-Nearest Neighbour 83.87% 70.60% 4 Neural Network Implementation This paper proposes a softmax neural net with 1 hidden layer and a fixed number of nodes (62) in the single hidden layer. The input features are 70 dimensional. The original problem asks to classify in 35 groups, corresponding to the number of letters plus the number of digits except for the 0 (which is considered to be the same as O and o ). Knowing that the data contains both upper and lower case letters, the approach of this paper proposes training a 62 output neural network with a posterior lookup table relating upper and lower case characters. This means that the Neural Network should be trained to distinguish the 62 different characters (62 outputs). We will use the multi-class cross entropy error function defined in equations 1 and 2, with a bias term for both the hidden and the output node. The Neural Network will be trained using a classic backpropagation algorithm. J(w) = m 3 i=1 k=1 out k (x (i) ) = ) y (i) k (out log k (x (i) ) exp(a out,k) j exp(a out,j) (1) (2) The minimization in the backpropagation algorithm implemented uses a 5-level depth 9- points long line search. For timing constraints the neural network was trained for a limited amount of iterations (15 iterations for each of the datapoints) and not until convergence. 5 Complete System Figure 3 shows a block diagram of the different steps in the recognition scheme. 6 Results In this section we present the results and a short analysis of the error committed.
6 Figure 3: Block diagram of the complete system Table 2: Algorithm comparison for Cross-Validation and UDNC Features Algorithm Max Recognition Rate Mean Recognition Rate 1-Nearest Neighbour 71.57% 67.71% 1-Layer Neural Network 75.81% 72.99% 6.1 Performance The results reported are related to considering 7 writers in the database as training and the rest (4) as testing. We performed 4 cross-validations (taking 4 different sets of 7 writers as training). Table 2 shows the performance of the Neural Network implementation versus the 1-Nearest Neighbour (baseline). 6.2 Error Analysis In order to see where does the Neural Network commit mistakes we build a confusion matrix as we can see in figure 4. The values of this matrix correspond to the best performing classification using Neural Networks. It is funny to see that there are several errors, specially confusing characters that, indeed, look alike. The most confused pairs can easily be found looking for the maximum of the confusion matrix. Those are the following: a-q, 9-q, r-t, r-y and 2-z. Figure 4: Confusion Matrix and a pairs of values where the Neural Network does not correctly recognize.
7 7 Conclusion and Future Work In this paper we have reviewed different kinds of features to recognize handwritten characters and proposed novel approach: the Uniform Differential Normalized Coordinates. These features outperform regular features for character recognition. The Uniform Differential Normalized Coordinates have been used to recognize the characters from a public database using standard Neural Networks. The results improve the baseline of the 1-Nearest Neighbour algorithm. Several other variants of this work could be reviewed more deeply. Future work on this project would include cross validation to check which is the optimal number of hidden neurons, comparison with other kinds of algorithms and iteration of neural networks until convergence instead of a certain number of iterations. Acknowledgments This work has been funded by the program of graduate fellowships of the Departamento de Becas, Investigacion y Universidades of Fundacion Caja Madrid. References [1] A. Asuncion and D. Newman. Uci machine learning repository. mlearn/mlrepository.html, [2] C. Bahlmann. Directional features in online handwriting recognition. Pattern Recognition, 39(1): , [3] M. Garris, C. Wilson, and J. Blue. Neural network-based systems for handprint ocr applications. IEEE Trans. Image Processing, (7): , [4] Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11): , [5] Y. LeCun and C. Cortes. The mnist database of handwritten digits. [6] D. Llorens, F. Prat, A. Marzal, and J. M. Vilar. The uji pen characters data set. [7] M. Parizeau, A. Lemieux, and C. Gagné. Character recognition experiments using unipen data. Proc. of 6th Int. Conf. on Document Analysis and Recognition (ICDAR), pages , [8] J. xiong Dong, A. Krzyzak, and C. Y. Suen. An improved handwritten chinese character recognition system using support vector machine. Pattern Recogn. Lett., 26(12): , 2005.
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals Chun Lei He, Ping Zhang, Jianxiong Dong, Ching Y. Suen, Tien D. Bui Centre for Pattern Recognition and Machine Intelligence,
Lecture 8 February 4
ICS273A: Machine Learning Winter 2008 Lecture 8 February 4 Scribe: Carlos Agell (Student) Lecturer: Deva Ramanan 8.1 Neural Nets 8.1.1 Logistic Regression Recall the logistic function: g(x) = 1 1 + e θt
A Content based Spam Filtering Using Optical Back Propagation Technique
A Content based Spam Filtering Using Optical Back Propagation Technique Sarab M. Hameed 1, Noor Alhuda J. Mohammed 2 Department of Computer Science, College of Science, University of Baghdad - Iraq ABSTRACT
Supporting Online Material for
www.sciencemag.org/cgi/content/full/313/5786/504/dc1 Supporting Online Material for Reducing the Dimensionality of Data with Neural Networks G. E. Hinton* and R. R. Salakhutdinov *To whom correspondence
Novelty Detection in image recognition using IRF Neural Networks properties
Novelty Detection in image recognition using IRF Neural Networks properties Philippe Smagghe, Jean-Luc Buessler, Jean-Philippe Urban Université de Haute-Alsace MIPS 4, rue des Frères Lumière, 68093 Mulhouse,
Simplified Machine Learning for CUDA. Umar Arshad @arshad_umar Arrayfire @arrayfire
Simplified Machine Learning for CUDA Umar Arshad @arshad_umar Arrayfire @arrayfire ArrayFire CUDA and OpenCL experts since 2007 Headquartered in Atlanta, GA In search for the best and the brightest Expert
Programming Exercise 3: Multi-class Classification and Neural Networks
Programming Exercise 3: Multi-class Classification and Neural Networks Machine Learning November 4, 2011 Introduction In this exercise, you will implement one-vs-all logistic regression and neural networks
Blood Vessel Classification into Arteries and Veins in Retinal Images
Blood Vessel Classification into Arteries and Veins in Retinal Images Claudia Kondermann and Daniel Kondermann a and Michelle Yan b a Interdisciplinary Center for Scientific Computing (IWR), University
Using Neural Networks to Create an Adaptive Character Recognition System
Using Neural Networks to Create an Adaptive Character Recognition System Alexander J. Faaborg Cornell University, Ithaca NY (May 14, 2002) Abstract A back-propagation neural network with one hidden layer
Efficient online learning of a non-negative sparse autoencoder
and Machine Learning. Bruges (Belgium), 28-30 April 2010, d-side publi., ISBN 2-93030-10-2. Efficient online learning of a non-negative sparse autoencoder Andre Lemme, R. Felix Reinhart and Jochen J. Steil
L25: Ensemble learning
L25: Ensemble learning Introduction Methods for constructing ensembles Combination strategies Stacked generalization Mixtures of experts Bagging Boosting CSCE 666 Pattern Analysis Ricardo Gutierrez-Osuna
Signature Segmentation from Machine Printed Documents using Conditional Random Field
2011 International Conference on Document Analysis and Recognition Signature Segmentation from Machine Printed Documents using Conditional Random Field Ranju Mandal Computer Vision and Pattern Recognition
Biometric Authentication using Online Signatures
Biometric Authentication using Online Signatures Alisher Kholmatov and Berrin Yanikoglu [email protected], [email protected] http://fens.sabanciuniv.edu Sabanci University, Tuzla, Istanbul,
Accurate and robust image superresolution by neural processing of local image representations
Accurate and robust image superresolution by neural processing of local image representations Carlos Miravet 1,2 and Francisco B. Rodríguez 1 1 Grupo de Neurocomputación Biológica (GNB), Escuela Politécnica
Neural Networks and Support Vector Machines
INF5390 - Kunstig intelligens Neural Networks and Support Vector Machines Roar Fjellheim INF5390-13 Neural Networks and SVM 1 Outline Neural networks Perceptrons Neural networks Support vector machines
Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승
Feed-Forward mapping networks KAIST 바이오및뇌공학과 정재승 How much energy do we need for brain functions? Information processing: Trade-off between energy consumption and wiring cost Trade-off between energy consumption
3 An Illustrative Example
Objectives An Illustrative Example Objectives - Theory and Examples -2 Problem Statement -2 Perceptron - Two-Input Case -4 Pattern Recognition Example -5 Hamming Network -8 Feedforward Layer -8 Recurrent
Analysis of kiva.com Microlending Service! Hoda Eydgahi Julia Ma Andy Bardagjy December 9, 2010 MAS.622j
Analysis of kiva.com Microlending Service! Hoda Eydgahi Julia Ma Andy Bardagjy December 9, 2010 MAS.622j What is Kiva? An organization that allows people to lend small amounts of money via the Internet
III. SEGMENTATION. A. Origin Segmentation
2012 International Conference on Frontiers in Handwriting Recognition Handwritten English Word Recognition based on Convolutional Neural Networks Aiquan Yuan, Gang Bai, Po Yang, Yanni Guo, Xinting Zhao
An Introduction to Data Mining. Big Data World. Related Fields and Disciplines. What is Data Mining? 2/12/2015
An Introduction to Data Mining for Wind Power Management Spring 2015 Big Data World Every minute: Google receives over 4 million search queries Facebook users share almost 2.5 million pieces of content
Towards better accuracy for Spam predictions
Towards better accuracy for Spam predictions Chengyan Zhao Department of Computer Science University of Toronto Toronto, Ontario, Canada M5S 2E4 [email protected] Abstract Spam identification is crucial
Lecture 6. Artificial Neural Networks
Lecture 6 Artificial Neural Networks 1 1 Artificial Neural Networks In this note we provide an overview of the key concepts that have led to the emergence of Artificial Neural Networks as a major paradigm
Unconstrained Handwritten Character Recognition Using Different Classification Strategies
Unconstrained Handwritten Character Recognition Using Different Classification Strategies Alessandro L. Koerich Department of Computer Science (PPGIA) Pontifical Catholic University of Paraná (PUCPR) Curitiba,
An Introduction to Neural Networks
An Introduction to Vincent Cheung Kevin Cannons Signal & Data Compression Laboratory Electrical & Computer Engineering University of Manitoba Winnipeg, Manitoba, Canada Advisor: Dr. W. Kinsner May 27,
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
1. Classification problems
Neural and Evolutionary Computing. Lab 1: Classification problems Machine Learning test data repository Weka data mining platform Introduction Scilab 1. Classification problems The main aim of a classification
6.2.8 Neural networks for data mining
6.2.8 Neural networks for data mining Walter Kosters 1 In many application areas neural networks are known to be valuable tools. This also holds for data mining. In this chapter we discuss the use of neural
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
Training Methods for Adaptive Boosting of Neural Networks for Character Recognition
Submission to NIPS*97, Category: Algorithms & Architectures, Preferred: Oral Training Methods for Adaptive Boosting of Neural Networks for Character Recognition Holger Schwenk Dept. IRO Université de Montréal
BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES
BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES 123 CHAPTER 7 BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES 7.1 Introduction Even though using SVM presents
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: [email protected]
Image Classification for Dogs and Cats
Image Classification for Dogs and Cats Bang Liu, Yan Liu Department of Electrical and Computer Engineering {bang3,yan10}@ualberta.ca Kai Zhou Department of Computing Science [email protected] Abstract
Support Vector Machines for Dynamic Biometric Handwriting Classification
Support Vector Machines for Dynamic Biometric Handwriting Classification Tobias Scheidat, Marcus Leich, Mark Alexander, and Claus Vielhauer Abstract Biometric user authentication is a recent topic in the
Making Sense of the Mayhem: Machine Learning and March Madness
Making Sense of the Mayhem: Machine Learning and March Madness Alex Tran and Adam Ginzberg Stanford University [email protected] [email protected] I. Introduction III. Model The goal of our research
Employer Health Insurance Premium Prediction Elliott Lui
Employer Health Insurance Premium Prediction Elliott Lui 1 Introduction The US spends 15.2% of its GDP on health care, more than any other country, and the cost of health insurance is rising faster than
Monotonicity Hints. Abstract
Monotonicity Hints Joseph Sill Computation and Neural Systems program California Institute of Technology email: [email protected] Yaser S. Abu-Mostafa EE and CS Deptartments California Institute of Technology
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
A simple application of Artificial Neural Network to cloud classification
A simple application of Artificial Neural Network to cloud classification Tianle Yuan For AOSC 630 (by Prof. Kalnay) Introduction to Pattern Recognition (PR) Example1: visual separation between the character
Impact of Boolean factorization as preprocessing methods for classification of Boolean data
Impact of Boolean factorization as preprocessing methods for classification of Boolean data Radim Belohlavek, Jan Outrata, Martin Trnecka Data Analysis and Modeling Lab (DAMOL) Dept. Computer Science,
CSCI567 Machine Learning (Fall 2014)
CSCI567 Machine Learning (Fall 2014) Drs. Sha & Liu {feisha,yanliu.cs}@usc.edu September 22, 2014 Drs. Sha & Liu ({feisha,yanliu.cs}@usc.edu) CSCI567 Machine Learning (Fall 2014) September 22, 2014 1 /
Taking Inverse Graphics Seriously
CSC2535: 2013 Advanced Machine Learning Taking Inverse Graphics Seriously Geoffrey Hinton Department of Computer Science University of Toronto The representation used by the neural nets that work best
High-Performance Signature Recognition Method using SVM
High-Performance Signature Recognition Method using SVM Saeid Fazli Research Institute of Modern Biological Techniques University of Zanjan Shima Pouyan Electrical Engineering Department University of
A fast multi-class SVM learning method for huge databases
www.ijcsi.org 544 A fast multi-class SVM learning method for huge databases Djeffal Abdelhamid 1, Babahenini Mohamed Chaouki 2 and Taleb-Ahmed Abdelmalik 3 1,2 Computer science department, LESIA Laboratory,
Transfer Learning for Latin and Chinese Characters with Deep Neural Networks
Transfer Learning for Latin and Chinese Characters with Deep Neural Networks Dan C. Cireşan IDSIA USI-SUPSI Manno, Switzerland, 6928 Email: [email protected] Ueli Meier IDSIA USI-SUPSI Manno, Switzerland, 6928
LVQ Plug-In Algorithm for SQL Server
LVQ Plug-In Algorithm for SQL Server Licínia Pedro Monteiro Instituto Superior Técnico [email protected] I. Executive Summary In this Resume we describe a new functionality implemented
A Simple Introduction to Support Vector Machines
A Simple Introduction to Support Vector Machines Martin Law Lecture for CSE 802 Department of Computer Science and Engineering Michigan State University Outline A brief history of SVM Large-margin linear
Chapter 6. The stacking ensemble approach
82 This chapter proposes the stacking ensemble approach for combining different data mining classifiers to get better performance. Other combination techniques like voting, bagging etc are also described
Classification algorithm in Data mining: An Overview
Classification algorithm in Data mining: An Overview S.Neelamegam #1, Dr.E.Ramaraj *2 #1 M.phil Scholar, Department of Computer Science and Engineering, Alagappa University, Karaikudi. *2 Professor, Department
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
Self Organizing Maps: Fundamentals
Self Organizing Maps: Fundamentals Introduction to Neural Networks : Lecture 16 John A. Bullinaria, 2004 1. What is a Self Organizing Map? 2. Topographic Maps 3. Setting up a Self Organizing Map 4. Kohonen
New Ensemble Combination Scheme
New Ensemble Combination Scheme Namhyoung Kim, Youngdoo Son, and Jaewook Lee, Member, IEEE Abstract Recently many statistical learning techniques are successfully developed and used in several areas However,
TOWARDS SIMPLE, EASY TO UNDERSTAND, AN INTERACTIVE DECISION TREE ALGORITHM
TOWARDS SIMPLE, EASY TO UNDERSTAND, AN INTERACTIVE DECISION TREE ALGORITHM Thanh-Nghi Do College of Information Technology, Cantho University 1 Ly Tu Trong Street, Ninh Kieu District Cantho City, Vietnam
A Simple Feature Extraction Technique of a Pattern By Hopfield Network
A Simple Feature Extraction Technique of a Pattern By Hopfield Network A.Nag!, S. Biswas *, D. Sarkar *, P.P. Sarkar *, B. Gupta **! Academy of Technology, Hoogly - 722 *USIC, University of Kalyani, Kalyani
SVM Ensemble Model for Investment Prediction
19 SVM Ensemble Model for Investment Prediction Chandra J, Assistant Professor, Department of Computer Science, Christ University, Bangalore Siji T. Mathew, Research Scholar, Christ University, Dept of
Low-resolution Character Recognition by Video-based Super-resolution
2009 10th International Conference on Document Analysis and Recognition Low-resolution Character Recognition by Video-based Super-resolution Ataru Ohkura 1, Daisuke Deguchi 1, Tomokazu Takahashi 2, Ichiro
SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS
UDC: 004.8 Original scientific paper SELECTING NEURAL NETWORK ARCHITECTURE FOR INVESTMENT PROFITABILITY PREDICTIONS Tonimir Kišasondi, Alen Lovren i University of Zagreb, Faculty of Organization and Informatics,
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
Artificial Neural Network, Decision Tree and Statistical Techniques Applied for Designing and Developing E-mail Classifier
International Journal of Recent Technology and Engineering (IJRTE) ISSN: 2277-3878, Volume-1, Issue-6, January 2013 Artificial Neural Network, Decision Tree and Statistical Techniques Applied for Designing
Class #6: Non-linear classification. ML4Bio 2012 February 17 th, 2012 Quaid Morris
Class #6: Non-linear classification ML4Bio 2012 February 17 th, 2012 Quaid Morris 1 Module #: Title of Module 2 Review Overview Linear separability Non-linear classification Linear Support Vector Machines
Neural Network based Vehicle Classification for Intelligent Traffic Control
Neural Network based Vehicle Classification for Intelligent Traffic Control Saeid Fazli 1, Shahram Mohammadi 2, Morteza Rahmani 3 1,2,3 Electrical Engineering Department, Zanjan University, Zanjan, IRAN
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS 1 AND ALGORITHMS Chiara Renso KDD-LAB ISTI- CNR, Pisa, Italy WHAT IS CLUSTER ANALYSIS? Finding groups of objects such that the objects in a group will be similar
Identity authentication using improved online signature verification method
Pattern Recognition Letters 26 (2005) 2400 2408 www.elsevier.com/locate/patrec Identity authentication using improved online signature verification method Alisher Kholmatov, Berrin Yanikoglu * Sabanci
Applied Mathematical Sciences, Vol. 7, 2013, no. 112, 5591-5597 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2013.
Applied Mathematical Sciences, Vol. 7, 2013, no. 112, 5591-5597 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2013.38457 Accuracy Rate of Predictive Models in Credit Screening Anirut Suebsing
Introduction to Machine Learning Using Python. Vikram Kamath
Introduction to Machine Learning Using Python Vikram Kamath Contents: 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. Introduction/Definition Where and Why ML is used Types of Learning Supervised Learning Linear Regression
Character Image Patterns as Big Data
22 International Conference on Frontiers in Handwriting Recognition Character Image Patterns as Big Data Seiichi Uchida, Ryosuke Ishida, Akira Yoshida, Wenjie Cai, Yaokai Feng Kyushu University, Fukuoka,
Support Vector Machine (SVM)
Support Vector Machine (SVM) CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
Visualization of large data sets using MDS combined with LVQ.
Visualization of large data sets using MDS combined with LVQ. Antoine Naud and Włodzisław Duch Department of Informatics, Nicholas Copernicus University, Grudziądzka 5, 87-100 Toruń, Poland. www.phys.uni.torun.pl/kmk
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING)
ARTIFICIAL INTELLIGENCE (CSCU9YE) LECTURE 6: MACHINE LEARNING 2: UNSUPERVISED LEARNING (CLUSTERING) Gabriela Ochoa http://www.cs.stir.ac.uk/~goc/ OUTLINE Preliminaries Classification and Clustering Applications
Open Access Research on Application of Neural Network in Computer Network Security Evaluation. Shujuan Jin *
Send Orders for Reprints to [email protected] 766 The Open Electrical & Electronic Engineering Journal, 2014, 8, 766-771 Open Access Research on Application of Neural Network in Computer Network
L13: cross-validation
Resampling methods Cross validation Bootstrap L13: cross-validation Bias and variance estimation with the Bootstrap Three-way data partitioning CSCE 666 Pattern Analysis Ricardo Gutierrez-Osuna CSE@TAMU
High-dimensional labeled data analysis with Gabriel graphs
High-dimensional labeled data analysis with Gabriel graphs Michaël Aupetit CEA - DAM Département Analyse Surveillance Environnement BP 12-91680 - Bruyères-Le-Châtel, France Abstract. We propose the use
Non-negative Matrix Factorization (NMF) in Semi-supervised Learning Reducing Dimension and Maintaining Meaning
Non-negative Matrix Factorization (NMF) in Semi-supervised Learning Reducing Dimension and Maintaining Meaning SAMSI 10 May 2013 Outline Introduction to NMF Applications Motivations NMF as a middle step
ADVANCED MACHINE LEARNING. Introduction
1 1 Introduction Lecturer: Prof. Aude Billard ([email protected]) Teaching Assistants: Guillaume de Chambrier, Nadia Figueroa, Denys Lamotte, Nicola Sommer 2 2 Course Format Alternate between: Lectures
Object Recognition and Template Matching
Object Recognition and Template Matching Template Matching A template is a small image (sub-image) The goal is to find occurrences of this template in a larger image That is, you want to find matches of
Supervised and unsupervised learning - 1
Chapter 3 Supervised and unsupervised learning - 1 3.1 Introduction The science of learning plays a key role in the field of statistics, data mining, artificial intelligence, intersecting with areas in
Data Mining Techniques
15.564 Information Technology I Business Intelligence Outline Operational vs. Decision Support Systems What is Data Mining? Overview of Data Mining Techniques Overview of Data Mining Process Data Warehouses
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
Research on Clustering Analysis of Big Data Yuan Yuanming 1, 2, a, Wu Chanle 1, 2
Advanced Engineering Forum Vols. 6-7 (2012) pp 82-87 Online: 2012-09-26 (2012) Trans Tech Publications, Switzerland doi:10.4028/www.scientific.net/aef.6-7.82 Research on Clustering Analysis of Big Data
Handwritten Digit Recognition with a Back-Propagation Network
396 Le Cun, Boser, Denker, Henderson, Howard, Hubbard and Jackel Handwritten Digit Recognition with a Back-Propagation Network Y. Le Cun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard,
AUTOMATION OF ENERGY DEMAND FORECASTING. Sanzad Siddique, B.S.
AUTOMATION OF ENERGY DEMAND FORECASTING by Sanzad Siddique, B.S. A Thesis submitted to the Faculty of the Graduate School, Marquette University, in Partial Fulfillment of the Requirements for the Degree
degrees of freedom and are able to adapt to the task they are supposed to do [Gupta].
1.3 Neural Networks 19 Neural Networks are large structured systems of equations. These systems have many degrees of freedom and are able to adapt to the task they are supposed to do [Gupta]. Two very
Sub-class Error-Correcting Output Codes
Sub-class Error-Correcting Output Codes Sergio Escalera, Oriol Pujol and Petia Radeva Computer Vision Center, Campus UAB, Edifici O, 08193, Bellaterra, Spain. Dept. Matemàtica Aplicada i Anàlisi, Universitat
Stochastic Pooling for Regularization of Deep Convolutional Neural Networks
Stochastic Pooling for Regularization of Deep Convolutional Neural Networks Matthew D. Zeiler Department of Computer Science Courant Institute, New York University [email protected] Rob Fergus Department
Introduction to Pattern Recognition
Introduction to Pattern Recognition Selim Aksoy Department of Computer Engineering Bilkent University [email protected] CS 551, Spring 2009 CS 551, Spring 2009 c 2009, Selim Aksoy (Bilkent University)
DIAGONAL BASED FEATURE EXTRACTION FOR HANDWRITTEN ALPHABETS RECOGNITION SYSTEM USING NEURAL NETWORK
DIAGONAL BASED FEATURE EXTRACTION FOR HANDWRITTEN ALPHABETS RECOGNITION SYSTEM USING NEURAL NETWORK J.Pradeep 1, E.Srinivasan 2 and S.Himavathi 3 1,2 Department of ECE, Pondicherry College Engineering,
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
Using Lexical Similarity in Handwritten Word Recognition
Using Lexical Similarity in Handwritten Word Recognition Jaehwa Park and Venu Govindaraju Center of Excellence for Document Analysis and Recognition (CEDAR) Department of Computer Science and Engineering
Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data
Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data Jun Wang Department of Mechanical and Automation Engineering The Chinese University of Hong Kong Shatin, New Territories,
NEURAL NETWORKS A Comprehensive Foundation
NEURAL NETWORKS A Comprehensive Foundation Second Edition Simon Haykin McMaster University Hamilton, Ontario, Canada Prentice Hall Prentice Hall Upper Saddle River; New Jersey 07458 Preface xii Acknowledgments
Personalized Hierarchical Clustering
Personalized Hierarchical Clustering Korinna Bade, Andreas Nürnberger Faculty of Computer Science, Otto-von-Guericke-University Magdeburg, D-39106 Magdeburg, Germany {kbade,nuernb}@iws.cs.uni-magdeburg.de
Reconstructing Self Organizing Maps as Spider Graphs for better visual interpretation of large unstructured datasets
Reconstructing Self Organizing Maps as Spider Graphs for better visual interpretation of large unstructured datasets Aaditya Prakash, Infosys Limited [email protected] Abstract--Self-Organizing
High-accuracy ultrasound target localization for hand-eye calibration between optical tracking systems and three-dimensional ultrasound
High-accuracy ultrasound target localization for hand-eye calibration between optical tracking systems and three-dimensional ultrasound Ralf Bruder 1, Florian Griese 2, Floris Ernst 1, Achim Schweikard
THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS
THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS O.U. Sezerman 1, R. Islamaj 2, E. Alpaydin 2 1 Laborotory of Computational Biology, Sabancı University, Istanbul, Turkey. 2 Computer Engineering
Chapter 7. Feature Selection. 7.1 Introduction
Chapter 7 Feature Selection Feature selection is not used in the system classification experiments, which will be discussed in Chapter 8 and 9. However, as an autonomous system, OMEGA includes feature
