Enhanced MLP Input-Output Mapping for Degraded Pattern Recognition

Similar documents
Figure 1. Artificial Neural Network structure. B. Spiking Neural Networks Spiking Neural networks (SNNs) fall into the third generation of neural netw

An Hybrid MLP-SVM Handwritten Digit Recognizer

A comparative study of different feature sets for recognition of handwritten Arabic numerals using a Multi Layer Perceptron

Sonia Sharma ECE Department, University Institute of Engineering and Technology, MDU, Rohtak, India. Fig.1.Neuron and its connection

DIAGNOSIS OF STATOR FAULT IN ASYNCHRONOUS MACHINE USING SOFT COMPUTING METHODS

Neural Filters: MLP VIS-A-VIS RBF Network

Contents 1 Introduction Optical Character Recognition Systems Soft Computing Techniques for Optical Character Recognition Systems

MINE 432 Industrial Automation and Robotics

A Robust Footprint Detection Using Color Images and Neural Networks

Segmentation of Fingerprint Images

IDENTIFICATION OF POWER QUALITY PROBLEMS IN IEEE BUS SYSTEM BY USING NEURAL NETWORKS

Detection of License Plates of Vehicles

Industrial computer vision using undefined feature extraction

10mW CMOS Retina and Classifier for Handheld, 1000Images/s Optical Character Recognition System


NEURALNETWORK BASED CLASSIFICATION OF LASER-DOPPLER FLOWMETRY SIGNALS

Open Access An Improved Character Recognition Algorithm for License Plate Based on BP Neural Network

Neural Network with Median Filter for Image Noise Reduction

Contrast adaptive binarization of low quality document images

FACE RECOGNITION USING NEURAL NETWORKS

Abstract. Most OCR systems decompose the process into several stages:

Artificial Neural Networks. Artificial Intelligence Santa Clara, 2016

Automatic Licenses Plate Recognition System

A Chinese License Plate Recognition System

Application of Classifier Integration Model to Disturbance Classification in Electric Signals

Neural Network Classifier and Filtering for EEG Detection in Brain-Computer Interface Device

Real Time ALPR for Vehicle Identification Using Neural Network

Identification of Cardiac Arrhythmias using ECG

Handwritten Character Recognition using Different Kernel based SVM Classifier and MLP Neural Network (A COMPARISON)

NEURAL NETWORK DEMODULATOR FOR QUADRATURE AMPLITUDE MODULATION (QAM)

An Improved Bernsen Algorithm Approaches For License Plate Recognition

SIMULATION-BASED MODEL CONTROL USING STATIC HAND GESTURES IN MATLAB

Application of Multi Layer Perceptron (MLP) for Shower Size Prediction

Research on Application of Conjoint Neural Networks in Vehicle License Plate Recognition

An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods

The Use of Neural Network to Recognize the Parts of the Computer Motherboard

Adaptive Multi-layer Neural Network Receiver Architectures for Pattern Classification of Respective Wavelet Images

Preprocessing and Segregating Offline Gujarati Handwritten Datasheet for Character Recognition

Introduction to Machine Learning

Classification Experiments for Number Plate Recognition Data Set Using Weka

Neural Labyrinth Robot Finding the Best Way in a Connectionist Fashion

Transactions on Information and Communications Technologies vol 1, 1993 WIT Press, ISSN

Image binarization techniques for degraded document images: A review

Current Harmonic Estimation in Power Transmission Lines Using Multi-layer Perceptron Learning Strategies

An Algorithm for Fingerprint Image Postprocessing

Evolving High-Dimensional, Adaptive Camera-Based Speed Sensors

Implementation of Text to Speech Conversion

AN IMPROVED NEURAL NETWORK-BASED DECODER SCHEME FOR SYSTEMATIC CONVOLUTIONAL CODE. A Thesis by. Andrew J. Zerngast

COMPARATIVE ANALYSIS OF ACCURACY ON MISSING DATA USING MLP AND RBF METHOD V.B. Kamble 1, S.N. Deshmukh 2 1

PHASE PRESERVING DENOISING AND BINARIZATION OF ANCIENT DOCUMENT IMAGE

Study and Analysis of various preprocessing approaches to enhance Offline Handwritten Gujarati Numerals for feature extraction

IDENTIFICATION OF SIGNATURES TRANSMITTED OVER RAYLEIGH FADING CHANNEL BY USING HMM AND RLE

258 IEEE TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS PART B: CYBERNETICS, VOL. 33, NO. 2, APRIL 2003

A Quantitative Comparison of Different MLP Activation Functions in Classification

ARTIFICIAL NEURAL NETWORK BASED CLASSIFICATION FOR MONOBLOCK CENTRIFUGAL PUMP USING WAVELET ANALYSIS

Live Hand Gesture Recognition using an Android Device

Automatic Locating the Centromere on Human Chromosome Pictures

License Plate Localisation based on Morphological Operations

Paper Sobel Operated Edge Detection Scheme using Image Processing for Detection of Metal Cracks

Modern spectral analysis of non-stationary signals in power electronics

Prediction of Cluster System Load Using Artificial Neural Networks

Radiated EMI Recognition and Identification from PCB Configuration Using Neural Network

A COMPARISON OF ARTIFICIAL NEURAL NETWORKS AND OTHER STATISTICAL METHODS FOR ROTATING MACHINE

Improvement of Classical Wavelet Network over ANN in Image Compression

Chapter - 7. Adaptive Channel Equalization

CHAPTER 6 BACK PROPAGATED ARTIFICIAL NEURAL NETWORK TRAINED ARHF

Estimation of Ground Enhancing Compound Performance Using Artificial Neural Network

VEHICLE LICENSE PLATE DETECTION ALGORITHM BASED ON STATISTICAL CHARACTERISTICS IN HSI COLOR MODEL

Keyword: Morphological operation, template matching, license plate localization, character recognition.

LabVIEW based Intelligent Frontal & Non- Frontal Face Recognition System

Background Pixel Classification for Motion Detection in Video Image Sequences

USING EMBEDDED PROCESSORS IN HARDWARE MODELS OF ARTIFICIAL NEURAL NETWORKS

J. C. Brégains (Student Member, IEEE), and F. Ares (Senior Member, IEEE).

Number Plate Recognition Using Segmentation

Classification in Image processing: A Survey

Colour Recognition in Images Using Neural Networks

Image Extraction using Image Mining Technique

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES

Robust Hand Gesture Recognition for Robotic Hand Control

Student: Nizar Cherkaoui. Advisor: Dr. Chia-Ling Tsai (Computer Science Dept.) Advisor: Dr. Eric Muller (Biology Dept.)

A Novel Morphological Method for Detection and Recognition of Vehicle License Plates

Effective and Efficient Fingerprint Image Postprocessing

Application of Artificial Neural Networks System for Synthesis of Phased Cylindrical Arc Antenna Arrays

Libyan Licenses Plate Recognition Using Template Matching Method

Publication P IEEE. Reprinted with permission.

Number Plate Detection with a Multi-Convolutional Neural Network Approach with Optical Character Recognition for Mobile Devices

The Basic Kak Neural Network with Complex Inputs

INDIAN VEHICLE LICENSE PLATE EXTRACTION AND SEGMENTATION

Feature analysis of EEG signals using SOM

IJITKMI Volume 7 Number 2 Jan June 2014 pp (ISSN ) Impact of attribute selection on the accuracy of Multilayer Perceptron

Analysis of Learning Paradigms and Prediction Accuracy using Artificial Neural Network Models

A Novel Fuzzy Neural Network Based Distance Relaying Scheme

Artificial Neural Network Engine: Parallel and Parameterized Architecture Implemented in FPGA

Impulse Noise Removal Based on Artificial Neural Network Classification with Weighted Median Filter

MULTIPLE CLASSIFIERS FOR ELECTRONIC NOSE DATA

Blind Dereverberation of Single-Channel Speech Signals Using an ICA-Based Generative Model

Developing a New Color Model for Image Analysis and Processing

Automatics Vehicle License Plate Recognition using MATLAB

Harmonic detection by using different artificial neural network topologies

Architecture design for Adaptive Noise Cancellation

Transcription:

Enhanced MLP Input-Output Mapping for Degraded Pattern Recognition Shigueo Nomura and José Ricardo Gonçalves Manzan Faculty of Electrical Engineering, Federal University of Uberlândia, Uberlândia, MG, Brasil Abstract This work proposes a set of approaches for improving the multilayer perceptron (MLP) performance on degraded pattern input-output mapping process First, differently to the classical one-per-class approach, our strategy calculates Euclidean distances between the MLP output and target vectors Second, our approach adopts orthogonal bipolar vectors (OBVs) as target values taking advantages of larger Euclidean distance provided by these vectors rather than conventional ones The proposed approaches were applied to MLP training and test in classifying very degraded patterns as input data Experimental results with classical approaches in parallel to the proposed ones are presented for MLP performance comparison purposes The improved MLP with our proposed approaches provided an increase of 93 % on degraded pattern recognition rate Keywords: Artificial Neural Networks, Degraded Pattern, Euclidean Distance, Multilayer Perceptron, Pattern Recognition, Target Vector 1 Introduction Multilayer perceptron (MLP) has been successfully applied to pattern recognition [1][2][3] tasks As a classification system, MLP requires a good approach for analyzing degraded image data, extracting features from these data, generating a set of relevant information, and improving its performance Many efforts have attempted to develop a good method followed by feature extraction systems Surprisingly, it is quite difficult to find investigations focusing on output space for the input-output mapping process of MLPs as shown in Fig 1 Basically, approaches for the MLP multiclass categorization tasks such as one-per-class based on conventional binary or bipolar target vectors with their dimensions equal to the number of classes are classical This work proposes to break away from such classical treatment of the MLP input-output mapping process The main objective of this work is to show that the proposed approaches can considerably improve the degraded pattern recognition performance by MLPs In summary, the approaches adopt the Euclidean distancebased MLP multiclass categorization and the use of orthogonal bipolar vectors (OBVs) as expectation values for learning It is known that MLP performance using bipolar target vectors is already better than performance with binary ones [4], but no other authors work exists, to our knowledge, adopting the Euclidean distance between OBVs and output vectors 2 Usual Approaches on Multiclass Categorization Connectionist algorithms are more difficult to apply to multiclass categorization problems [5] Multiclass categorization problems [5] correspond to tasks of finding an approximate definition for an unknown function f(x) given training examples of the form x i, f(x i ) The unknown function f often takes values from a discrete set of classes c 1, c 2,, c k We can distinguish two approaches to handle these multiclass categorization tasks as follows: We have one-per-class approach when the individual functions f 1, f 2,, f k are learned one for each class To assign a new case p to one of these classes, each of individual function f i is evaluated on p, and the case p is assigned the class j corresponding to the function f j that returns the highest activation [6] This categorization approach is standard for conventional target vectors Distributed output code is an alternative approach pioneered by Sejnowski and Rosenberg [7] in their widelyknown NETtalk system In this approach, each class is assigned a unique binary string of length n; these strings refer to target vectors in MLP Then n binary functions are learned, one for each bit position in these binary strings During training for an example from class i, the desired outputs of these n binary functions are specified by the target vector for class i With Artificial Neural Networks (ANN), these n functions can be implemented by the n output units of a single network A new case p is classified by evaluating each of the n binary functions to generate an n-bit string s This string is then compared to each of the k target vectors, and p is assigned to the class whose target vector is closest, according to some distance measure, to the generated string s

MLP M-dimensional Input space N-dimensional Output space Fig 1: Input-output mapping process of an MLP 3 Motivation An MLP model is trained to learn the nonlinear relationship between M-dimensional input space and N- dimensional output space as shown in Fig 1 In this way, an incoming unknown input M-dimensional vector is transformed into a N-dimensional one and it can be placed in any point from the space The motivation of this work is to improve the MLP multiclass categorization performance as pattern recognition process by transforming an input vector representing a degraded pattern into an output vector representing a class The increase of the generalization ability of MLP models leads to their recognition performance improvement One goal is to find a topology design strategy to provide a network with a higher ability to generalize Many traditional investigations have concentrated on improving the representation of input vectors On the contrary, this work has concentrated on the alternative representation of the desired outputs (OBVs as target vectors) to adjust weights that minimize error during the supervised training process A purpose is to influence on the multidimensional error-performance surface [8] that is constructed in the supervised training Since the categorization based on Euclidean distance adopted in this work depends on the distance between vectors to classify unknown input data, it is expected that the MLP performance improves with the proposed approaches The orthogonality of new target vectors can provide larger output space rather than conventional vectors Figure 2 presents high interference zones between pattern categorization subspaces due to the conventional target vectors On the other hand, we can verify low interference zones between pattern categorization subspaces in Fig 3 due to the use of new target vectors Our hypothesis is that the approaches can lead to enhanced MLP input-output mapping even when input is degraded and slightly different from the training examples 4 Work Proposal A set of approaches including the input-output mapping process of MLPs based on Euclidean distance measure adopting OBVs as new target vectors is proposed in this work According to Dekel and Singer [9], the objective in multiclass categorization problems is to learn a classifier that accurately assigns labels (target vectors) to instances (input vectors) where the set of labels is of finite cardinality and contains more than two elements In this MLP application for multiclass categorization, after every presentation of an instance V (input vector in M- dimensional space of Fig 1), the Euclidean distance of the output vector in N-dimensional space was calculated with all labels (target vectors in output space) for each of the classes The class whose representation by the label had the smallest Euclidean distance with the output vector would be chosen as the winner to assign to instance V In addition to the Euclidean distance-based approach, we propose the use of OBVs (defined in the next section) as target vectors 5 Definition of Vectors The following defined vectors represent target values used in our experiments for the MLP input-output mapping analysis 51 Conventional Bipolar Vectors (CBVs) The conventional bipolar vectors (CBVs) with n components for representing p th pattern in n patterns are defined by Eq (1): p 1 n p {}}{{}}{ V p = ( 1,, 1, 1, 1,, 1) T, (1) where V p is the CBV for representing the p th pattern, p = 1, 2,, n, n is the number of patterns or components In case of recognition problem for 10 digits, the digit 0 is defined as a 10 th digit by Eq (2): 9 {}}{ V 0 = ( 1,, 1, 1) T (2)

Input pattern 1 Target vector 1 Interferencezones Target vector 2 Input pattern 2 MLP Target vector 3 Input pattern m Categorization subspace Target vector n Fig 2: High interference zones between categorization subspaces in case of using conventional target vectors vector 1 Input pattern 1 Input pattern 2 Input pattern m MLP Categorization subspace vector 2 vector 3 vector n Interferencezones Fig 3: Low interference zones between categorization subspaces in case of using new target vectors 52 Orthogonal Bipolar Vectors (OBVs) The norm of an orthogonal bipolar vector (OBV) in an Euclidean space R n is given by Eq (3): U = x 12 + x 22 + + x n2 = n, (3) where U = (x 1, x 2,, x n ) T, x i represents a component +1 or 1 for i = 1, 2,, n, and n is the number of components The usual inner product [10] between two vectors U and V in an Euclidean space R n is defined by Eq (4): UV = x 1 y 1 + x 2 y 2 + + x n y n, (4) where V = (y 1, y 2,, y n ) T, and y i represents a component +1 or 1 for i = 1, 2,, n, and n is the number of components Vectors U and V are orthogonal (denoted by U V ) if and only if UV = 0 in Eq (5): U V UV = 0 (5) A simple algorithm [4][11] can be implemented to generate the OBVs with various numbers of components following the above conditions Table 1 presents examples of OBVs with 16 components, representing target values for 10 digits, generated by the mentioned algorithm 6 Experimental Procedure The modeling procedure of an MLP topology for the input-output mapping process of degraded pattern data as shown in Fig 4 is presented in this section The model is to experimentally evaluate the proposed approaches applied to the MLP in degraded pattern recognition tasks We have extracted input data from license plate photos automatically taken by traffic control systems of Uberlândia City in Brasil They were images with such problems as luminosity, contrast, focalization, resolution, and size, all of which required preprocessing able to extract relevant features for pattern recognition process The original preprocessing methods proposed in such previous works as adaptive contrast enhancement [12], adaptive thresholding [13], automatic segmentation and extraction of feature vectors [14] were used

Table 1: Examples of OBVs with 16 components Digit OBV 1 (1, -1, -1, 1, -1, 1, 1, -1, -1, 1, 1, -1, 1, -1, -1, 1) 2 (1, -1, -1, 1, -1, 1, 1, -1, 1, -1, -1, 1, -1, 1, 1, -1) 3 (1, -1, -1, 1, 1, -1, -1, 1, -1, 1, 1, -1, -1, 1, 1, -1) 4 (1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1, 1, -1, -1, 1) 5 (1, -1, 1, -1, -1, 1, -1, 1, -1, 1, -1, 1, 1, -1, 1, -1) 6 (1, -1, 1, -1, -1, 1, -1, 1, 1, -1, 1, -1, -1, 1, -1, 1) 7 (1, -1, 1, -1, 1, -1, 1, -1, -1, 1, -1, 1, -1, 1, -1, 1) 8 (1, -1, 1, -1, 1, -1, 1, -1, 1, -1, 1, -1, 1, -1, 1, -1) 9 (1, 1, -1, -1, -1, -1, 1, 1, -1, -1, 1, 1, 1, 1, -1, -1) 0 (1, 1, -1, -1, -1, -1, 1, 1, 1, 1, -1, -1, -1, -1, 1, 1) 61 Data Representation A two-dimensional set of pixels that represents a pattern is mapped onto an input vector as a set of input neurons For a 20 15 image of each segmented pattern, the top row of 20 pixels is associated with the first 20 neurons, the next row of 20 pixels is associated with the next 20 neurons, and so on So, segmented entities (20 15) are represented by feature vectors with 300 components, and each component in the vector should represent one pixel of the pattern (bipolar value +1) or one pixel of the image background (bipolar value 1) Figure 4 shows the 120-image training set representing digits as input data In this work, the adopted data representation is bipolar since the categorization may be improved if the input is represented in bipolar form and the bipolar sigmoid is used for the activation function [4] The reason is if one factor in the weight connection expression is the activation of the lower unit then units whose activations are null will not learn [4] 62 MLP Topology for Experiments The adopted multilayer neural network in our experiments consists of the architecture with one layer of hidden neurons There are several propositions how to determine the necessary number of hidden neurons for a given problem but they yield contradictory results and have no practical utility [15] Some heuristic rules [15] used to obtain the MLP topology for experiments are as follows: To reduce the number of hidden neurons when the network does not generalize, that is, the error of the output data during the training is small and the error in the test stage is large To increase the number of hidden neurons when the error during the training is large or all the weights for the connections between neurons are of the same order of magnitude In degraded images, the input data contain redundant information which can not be removed by a suitable coding These spurious data can be removed when the hidden layer contains less neurons than the input layer [15] In summary, without further information there is no foolproof method for setting the exact number of hidden neurons before training stage [4] Such usual strategies in ANN [8] as one parameter keeping and the variation of remaining parameters defined the appropriate topology of MLP Conventional experiments get adequate topology for classifying input digits (20 15) represented by the 300-dimensional feature vectors Our experimental MLP model consists of 300 neurons in the input layer The adequate number of neurons in the hidden layer is set according to each experiment The number of neurons in the output layer is defined by the target vector type or its size selected for each experiment 63 Training Stage The standard backpropagation algorithm [4] is used as the training algorithm of each MLP model Since all experimental target vectors are bipolar, the adopted activation function is the typical bipolar sigmoid [4], which has a range of ( 1, 1) Initial weights are generated as random values between 025 and 025 The learning rate parameter is set as 002 The stop criterion for the training algorithm is to require that the maximum value of the average squared error be equal to or less than the tolerance as indicated within each graph of Figs 5 7 Training data set is constituted by 120 pattern-images not belonging to the test data set It contains input patterns for the MLP model training to classify digits extracted from license plates into ten categories Each category is represented by 12 input patterns 64 Test Stage The classification rate is calculated by Eq (6):

Fig 4: A sample of 120 degraded patterns used as input data for the MLP training Classification rate (%) 724 722 720 718 716 718 717 717 717 720 720 720 721 3751 720 5496 714 712 711 712 2324 2864 710 708 706 103 149 278 341 443 643 1220 704 30 20 10 08 06 04 02 01 008 006 004 Tolerance (x10-3 ) Fig 5: MLP performance using classical one-per-class approach The line denotes the epoch quantity evolution

Classification rate (%) 801 799 797 795 797 798 799 800 796 796 63 93 795 795 126 793 791 43 22 27 32 38 38 30 25 20 17 11 07 05 Tolerance (x10-2 ) Fig 6: MLP performance using OBVs with 16 components as target vectors The line denotes the epoch quantity evolution Classification rate (%) 816 814 812 810 808 806 814 813 195 812 812 811 811 809 809 102 69 53 28 32 37 43 40 35 30 25 20 15 10 05 Tolerance (x10-2 ) Fig 7: MLP performance using OBVs with 64 components as target vectors The line denotes the epoch quantity evolution cr = N i c i N, (6) where cr is classification rate, N is number of test patterns, and c i is defined as c i = { 1 : pi = r i 0 : p i r i, (7) where p i is a classified pattern (output of the MLP model), and r i is the corresponding category (correct response) Test data set consisted of 1352 images containing the digits 0, 1,, 9 extracted from license plates after their preprocessing [12][13][14][16] A trained MLP model was applied to classify those extracted digits into ten categories 7 Experiments and Results The experiments consisted of using CBVs and OBVs defined in Section 5 as target vectors We have evaluated the influence of Euclidean distancebased approach on MLP training and classification performance improvement using various target vectors In other words, the experiments compared the proposed Euclidean distance-based approach with the classical oneper-class categorization approach in terms of influence on MLP performance An MLP model with a topology of 100 hidden neurons and 10 output neurons was trained and tested using CBVs for one-per-class approach Figure 5 shows the MLP classification performance applying one-per-class approach for the use of CBVs as target vectors To evaluate the influence of using OBVs as target vectors on the MLP input-output mapping process, we have performed the following tasks: MLP model training for OBVs as target vectors Different topologies according to the OBV sizes (16 or 64 components) were defined

Table 2: Relevant results from the graphs of Figs 5 7 Target vector type Categorization approach Maximum rate % CBV One-per-class 7210 OBV-16 Euclidean distance-based 8000 OBV-64 Euclidean distance-based 8140 Application of the Euclidean distance-based approach for the above trained MLP models The experimental results using OBVs with 16 and 64 components, are respectively shown in Figs 6 and 7 8 Discussion The graph in Fig 5 shows that the classification rate of 721% is provided by the MLP model trained for 3751 epochs with the one-per-class categorization approach Figures 6 7 show the results from the use of OBVs as target values Table 2 presents the highest classification rate per target vector type and categorization approach We can note that the classification rate was increased of 93% by using Euclidean distance-based approach and OBVs as target values Considering the advantage of adopting the Euclidean distance-based categorization approach, we can verify that the MLP performance improved more and more with the use of OBVs rather than conventional target vectors 9 Conclusion A set of approaches to improve the MLP performance on degraded pattern recognition searching for more generalized and well-trained model was proposed Basically, the Euclidean distance-based approach was combined to the use of OBVs as new target vectors for the MLP training and test with very degraded pattern data We compared the experimental results applying the proposed approaches with the results applying the classical oneper-class approach Comparison results showed that the proposed approaches considerably improved (increase of 93% on classification rate) the MLP performance on degraded pattern recognition tasks In summary, we realized that the proposed approaches led to higher classification rate of very degraded pattern data by MLPs compared to the classical one-per-class approach In terms of degraded pattern recognition performance, we have gotten an enhanced MLP input-output mapping with the proposed Euclidean distance-based approach using OBVs as new target vectors References [1] A R Webb and K D Copsey Statistical Pattern Recognition John Wiley & Sons, third edition, 2011 [2] J Chu, I Moon, and M Mun A real-time EMG pattern recognition system based on linear-nonlinear feature projection for a multifunction myoelectric hand IEEE Transactions on Biomedical Engineering, 53(11):2232 2239, Nov 2006 [3] Y X Zhang Artificial neural networks based on principal component analysis input selection for clinical pattern recognition analysis Talanta, 73(1):68 75, Aug 2007 [4] L V Fausett Fundamentals of Neural Networks: Architectures, Algorithms, and Applications Prentice Hall, Englewood Cliffs, NJ, 1994 [5] T G Dietterich and G Bakiri Solving multiclass learning problems via error-correcting output codes Journal of Artificial Intelligence Research, 2:263 286, 1995 [6] N J Nilsson Learning Machines McGraw-Hill, New York, 1965 [7] T J Sejnowski and C R Rosenberg Parallel networks that learn to pronounce english text Complex Systems, 1:145 168, 1987 [8] S Haykin Neural Networks: A Comprehensive Foundation Prentice Hall, New Jersey, second edition, 1999 [9] O Dekel and Y Singer Multiclass learning by probabilistic embeddings In Advances in Neural Information Processing Systems, volume 15, pages 945 952 MIT Press, 2002 [10] B Noble and J W Daniel Applied Linear Algebra Prentice Hall, Englewood Cliffs, New Jersey, second edition, 1977 [11] S Nomura, K Yamanaka, O Katai, H Kawakami, and T Shiose Improved MLP learning via orthogonal bipolar target vectors Journal of Advanced Computacional Intelligence and Intelligent Informatics, 9(6):580 589, Nov 2005 [12] S Nomura, K Yamanaka, O Katai, and H Kawakami A new method for degraded color image binarization based on adaptive lightning on grayscale versions IEICE Trans on Information and Systems, E87- D(4):1012 1020, 2004 [13] S Nomura and K Yamanaka New adaptive methods applied to binarization of printed word images, In N Younan, editor, Proceedings of the Fourth IASTED International Conference Signal and Image Processing, pages 288 293, Kauai, USA, 2002 [14] S Nomura, K Yamanaka, O Katai, H Kawakami, and T Shiose A novel adaptive morphological approach for segmenting characters in degraded images Pattern Recognition, 38:1961 1975, Nov 2005 [15] (2012) Neuronal networks: The network [Online] Available: http://wwwandreas-mielkede/nn-en-4html [16] S Nomura and K Yamanaka New adaptive approach based on mathematical morphology applied to character segmentation and code extraction from number plate images, In Proc of 6 th World Multi- Conference on Systemics, Cybernetics and Informatics, volume IX, Florida, USA, Jul 2002 Acknowledgment The authors would like to thank PROPP-UFU (project 72/2010), CAPES (MINTER UFU-IFTM), and FAU-UFU for supporting this work