chapter one introduction and literature reviewdocs.neu.edu.tr/library/6315504929/tez.pdf · chapter...

62
1 CHAPTER ONE INTRODUCTION AND LITERATURE REVIEW 1.1 Introduction Biometrics is the study of methods for measuring physical or behavioral traits of an individual that can be used for uniquely recognizing or verifying that individual‟s identity [1]. Physical biometrics based on face recognition, eye print, finger print, and voice recognition are common as they are easier to measure than behavioral characteristics such as handwriting and printing type. The research‟s on biometrics are taking a lot of interest from researchers and scientists due to its high importance in security and safety systems and the increasing needs for more and more secure and easy recognition methods. The use of cards with long difficult passwords represent a big challenge for a lot of people who face difficulties in remembering their bank accounts, credit cards, even there own computers passwords. The revolution which happened in the digital technology has encouraged researchers to make advanced steps in biometrics. The day life use of digital recognition systems based on biometrics became a reality. Biometrics are divided into two distinct methods, these are physical methods like face and eye, and behavioral methods like printing type or walk rhythm. Applications for biometrics are most common in security, medical, and robotics areas related to fingerprint, face, iris, and gait. These biometric areas have gained the most attention among the research community. The potential for using the ear‟s appearance as a means of personal identification was recognized and advocated as early as 1890 by the French criminologist Alphonse Bertillon [1]. Ear recognition has been mentioned and used recently in many researches. The interest in ear as a physical feature of human is due to its stability over the years. The ear of a human remain same for the duree of the life with no important changes. Ear images can be acquired in a similar manner to face images, and a number of researchers have suggested that the human ear is unique enough to each individual to allow practical use as a biometric [2]. Many researchers have used the two dimension ear recognition, other used 3D recognition for ear features. In fact,

Upload: others

Post on 30-Oct-2019

11 views

Category:

Documents


0 download

TRANSCRIPT

1

CHAPTER ONE

INTRODUCTION AND LITERATURE REVIEW

1.1 Introduction

Biometrics is the study of methods for measuring physical or behavioral traits of an

individual that can be used for uniquely recognizing or verifying that individual‟s

identity [1]. Physical biometrics based on face recognition, eye print, finger print, and

voice recognition are common as they are easier to measure than behavioral

characteristics such as handwriting and printing type.

The research‟s on biometrics are taking a lot of interest from researchers and

scientists due to its high importance in security and safety systems and the increasing

needs for more and more secure and easy recognition methods. The use of cards with

long difficult passwords represent a big challenge for a lot of people who face

difficulties in remembering their bank accounts, credit cards, even there own

computers passwords. The revolution which happened in the digital technology has

encouraged researchers to make advanced steps in biometrics. The day life use of

digital recognition systems based on biometrics became a reality.

Biometrics are divided into two distinct methods, these are physical methods like face

and eye, and behavioral methods like printing type or walk rhythm. Applications for

biometrics are most common in security, medical, and robotics areas related to

fingerprint, face, iris, and gait. These biometric areas have gained the most attention

among the research community.

The potential for using the ear‟s appearance as a means of personal identification was

recognized and advocated as early as 1890 by the French criminologist Alphonse

Bertillon [1]. Ear recognition has been mentioned and used recently in many

researches. The interest in ear as a physical feature of human is due to its stability

over the years. The ear of a human remain same for the duree of the life with no

important changes. Ear images can be acquired in a similar manner to face images,

and a number of researchers have suggested that the human ear is unique enough to

each individual to allow practical use as a biometric [2]. Many researchers have used

the two dimension ear recognition, other used 3D recognition for ear features. In fact,

2

the ear may already be used informally as a biometric. For example, the United States

Immigration and Naturalization Service (INS) has a form giving specifications for the

photograph that indicate that the right ear should be visible (figure 1.1) [2].

Figure 1.1 : INS Form-378 (6-92) Asking for the Right Ear to be Visible [2]

The ear biometric problem (as with all biometrics) is split between enrolment and

recognition. Enrolment is the discovery, localization and normalization of the ear

image, whilst recognition deals with identification of a subject by ears. Currently,

most enrolment for experiments is done manually, but in any real world test of ear

biometrics, enrolment is just as important a problem as recognition. Enrolment has not

seen the same amount of research as recognition but there is some development in the

area [3].

Artificial neural networks has become very famous and applied in many aspects of

science. It is using mathematical equations and formulas to simulate the function and

structure of the brain. Artificial neural networks and back propagation algorithm have

been used widely in biometrics. The use of artificial neural network in face and iris in

addition to fingerprint recognition has been introduced in many researches. Neural

networks are massively parallel computing systems consisting of a large number of

simple processors with many interconnections.

3

The principle of ANN is applied for approximating a function where they learn a

function by looking at examples of this function. Here the internal weights in the

ANN are slowly adjusted so as to produce the same output as in the examples.

Performance is improved over time by iteratively updating the weights in the network.

The hope is that when the ANN is shown a new set of input variables, it will give a

correct output [4]. The training of ANN to perform a task implies the use of sets of

inputs and targets. These sets are considered as examples for teaching the networks.

The internal connections of the network are modified iteratively until finding the

optimum values that approximate the best our input output relationship.

This thesis discusses the use of artificial neural networks in biometric systems. The

ear recognition has been specifically chosen to be performed using back propagation

neural networks. Ear pictures of 54 persons will be used in normal and noisy sets for

the training of ANN.

1.2 Literature Review

The use of ear has been introduced and mentioned since a long time. Many different

extraction and recognition methods have been mentioned and used in literature using

ear characteristics. Researchers are still discussing weather ear is unique or unique

enough to be used in biometrics. The use of ANN for ear recognition has been

introduced in [5]. The basics of using ear as biometric for person identification and

authentication are presented. In [6] a 3D image recognition based on special analysis

has been discussed. A ray transform has been used for ear image recognition in [3].

The use of ear biometrics for person identification based on geometrical structure has

been discussed in [7]. A novel ear enrolment technique using the image ray transform,

based upon an analogy to light rays. The transform is capable of highlighting tubular

structures such as the helix of the ear and spectacle frames and, by exploiting the

elliptical shape of the helix, can be used as the basis of a method for enrolment for ear

biometrics. A review of biometrics and ear recognition techniques has been discussed

in [8]. In [2] and [9] different 2D and 3D images processing techniques for ear

recognition has been discussed. The geometrical structures observed from pixel value

distances are used for the successful recognition of objects.

Spatial segmentation of 2D ear images for recognition purpose has been introduced in

[1]. The author in [10] proposed a geometric approach for ear recognition. A simple

4

two stage scale and rotation invariant geometric approach based on the concept of

maximum line was used. The use of wavelet transform based approach for ear

recognition has been introduced in [11]. A new banana wavelet transform was used

and its performance was compared with Gabor wavelet for ear recognition. an

experimental study to investigate the effect of time difference between image

acquisition for gallery and probe on the performance of ear recognition was also

carried out in this work. A study of computerized ear recognition methods has been

carried out in [12] and [13]. A discussion about ear recognition has been introduced in

[14]. Evaluation of ear recognition has been presented in [15] and [16]. A survey

about image features extractions for ear recognition has been applied by [17].

Biometric recognition using 3D ear shape was discussed in [18]. In [19], a multimodal

biometrics based on face and ear recognition has been discussed.

In [20], the authors discussed a new multi modal biometrics based on face and ear.

The research significance of multi-modal biometrics based on face and ear was

represented, Then some researchers‟ achievements and inadequacies. Different

researches and explorations were done in the paper. A new novel ear enrolment

technique using the image ray transform, based upon an analogy to light rays in [21].

The transform is capable of highlighting tubular structures such as the helix of the ear

and spectacle frames and, by exploiting the elliptical shape of the helix, can be used

as the basis of a method for enrolment for ear biometrics. The presented technique

achieved good results of success. Multimodal biometrics based on fingerprint and ear

biometrics for personal identification was presented in [22]. A novel technique of

edge interaction point detection (EIPD) was used to determine the ear features.

Fingerprint features are identified by line based connected component analysis and its

feature vectors are generated using EIPD system. Also, Neural network based back

propagation network is used for an identity verification system.

An experimental study to demonstrate the effect of the time difference in image

acquisition for gallery and probe on the performance of ear recognition was presented

in [23]. [24] Presented an online biometric authentication using ear contours acquired

from a robust peg free acquisition set up. Gaussian classifiers were used to first

segment the skin and non-skin areas in the ear images. Laplacian of Gaussian was

then used to compute edges of the skin areas; which helped to get ear images. A pixel

based feature extraction for ear biometrics method was proposed by [25].

5

Many other references has discussed the use of ear recognition for person

identification and/or person identity check using different methods. The use of neural

network for identification purpose has been proposed recently in many works. Its

simple structure and high performance in addition to its ease of use compared to the

other methods that need a lot of manual processing have put it in the first of

recognition methods. The work in this thesis proposes the use of ANNs for ear

biometrics based person identification.

6

CHAPTER TWO

BIOMETRICS

2.1 Overview

Confirming the identity of a person nowadays became a very important challenge

facing the society. There exist many identification methods which vary between high

secure and less secure methods. However, some applications and life fields impress

more need for highly secured identification such as military, health, police and banks.

Biometrics has been mentioned as one of the most accurate and secure identification

methods which are being given more and more attention by researchers in the field.

In the traditional classic identification methods, person is to memorize huge

combinations of letters and numbers in order to have access to some places or

applications. Unless these pass-codes or pass-words are correct, people can‟t have

instant access to their works or money.

According to Nilson report, in 2005, MasterCard, Visa, American Express and

Discover incurred US$ 1.14 billion in fraud losses. Also 20 to 50% of helpdesk calls

were for password resets and each password reset costs about US$70. Over more, in

2005, 9.3 million American citizens suffered from identity theft which costs a total of

US$ 54.4 billion [21].

The need for memorizing long pass-phrases and the possibility of these phrases to be

forgotten or even stolen encouraged the researchers more and more toward finding

easier, simple, and more secure identification methods. One of the most secure

methods of identification is biometrics.

Biometrics is the science that studies the behavior exhibited and physical properties in

order to be able to distinguish between people. It represents the study of automated

and semi-automated systems or methods for recognizing a person based on his

physical or behavioral characteristics. Biometrics can be divided into two categories,

identification systems and verification systems. Identification system tells “who are

you”, while verification system tell “are you the person you are claiming yourself to

be?” [10].

7

Biometrics has been used long time ago in the identification of persons; people in the

desert were using the physical and biological signs in the verification of a person

identity. Finger prints and signature has been used for long time as a verification

method hundreds of years ago. The development of science and digital world has

increased the ability of injecting these biometrics and others easily in all fields of the

life. Nowadays fingerprint identity recognition machines and voice recognition

machines are widely used even in very small workplaces.

Memory and token identification methods are implemented to judge if the specified

user should have an access to a specified resource. Therefore an exact identification of

a person is not necessary and indeed is not always performed. It is possible that a

group of people has the same token or know the same password. Biometric

identification methods start with proper identification of a person and only after that,

the proper rights are assigned. Thus the main difference between classic methods and

biometrics is that biometric properties cannot be „borrowed‟ so people cannot - in the

way as simple as giving a token or telling the password - propagate their rights to

others. It obviously increases security of the system but sometimes may cause

problem[22].

First stage in each biometric process is collecting a set of „samples‟ from every user

who should be identified by the system. A sample is a set of biometric data measured

for a person in a single measurement. The biometric data may be a different kind of

psycho-physiological measurements. Next stage in most methods is creating a

„template‟ for each user based on previously collected samples. A template is a kind

of mean from all samples collected for this user. The process of creating a template is

called an „enrolment‟ of the user.

2.2 Characteristics of biometrics

In the opinion of [22 and 23] any human physiological or behavioral characteristic

could be a biometrics just if it has the following desirable properties:

1- Universality, which means that every person should have the

characteristic, if some people doesn‟t have that property then it can‟t be

considered as biometrics.

8

2- Uniqueness, which indicates that no two persons should be the same in

terms of the characteristic, otherwise.

3- Permanence, which means that the characteristic should be invariant with

time.

4- Collectability, which indicates that the characteristic can be measured

quantitatively.

The author in[14] also has mentioned three other conditions which must be achieved

in order to consider a property as to be biometrics, these conditions are:

1- Performance, which refers to the achievable identification accuracy, the

resource requirements to achieve acceptable identification accuracy, and the

working or environmental factors that affect the identification accuracy.

2- Acceptability, which indicates to what extent people are willing to accept the

biometric system.

3- Circumvention, which refers to how easy it is to fool the system by fraudulent

techniques

A biometric feature is a specific attribute of a person that can be measured with some

precision. There are a lot of different biometric features that can be measured[22].

2.3 Evaluation of biometric identification

Measurement of biological quantities is always to some degree imprecise and

therefore is producing different values for the same quantity measured. These errors

are an instant part of every biometric method and the main problem of that kind of

identification is to elaborate algorithms that sufficiently deal with these imprecise

data.

There are two kinds of tests when considering authorization (two class) system:

• Genuine test – when a sample is given with correct identification

information (login). In another words „the identified person is telling the

truth‟. In such case the rate of improper rejections may be measured. This

9

measure is often called a False Rejection Rate (FRR) or False Non-Match

Rate.

• Impostor test – when a sample is given with incorrect login. In another

words „the identified person is lying‟. Now a rate of improper acceptances

may be measured. This measure is called a False Acceptance Rate (FAR) or

False Match Rate [22].

The comparison between the two measures seems to be a good approach in order to

evaluate the identification method. However, still the two measures are related to each

other and decreasing one of them will increase the other.

2.4 Different Biometrics Methods

Most of human body parts has been used or can be used as biometrics, that due to the

fact that the features of each person‟s parts are unique and can‟t be found in the other

persons‟ bodies. The biometrics is divided into two types: the first uses the measure of

fingerprint, iris, eye retina, face, palm teeth, ear, and even smell for some measures.

The second uses the measure of humans‟ behavior patterns such as his way of

speaking, walking, shape of signature, and the hand writing.

Though behavior biometrics is less expensive and less dangerous for the user, physic-

logical characteristics offer highly exact identification of a person. Nevertheless, all

two types provide high level of identification than passwords and cards.

2.5 Physiological types of biometrics

Physiological systems are considered to be more reliable as individual features of a

person, that are used by these systems, do not change by influence of psycho-

emotional state.

2.5.1 Finger Print

Fingerprints are considered being one of the oldest and popular among other bio-

metric technologies. First uses of fingerprints instead of signatures were reported in

19th century. The milestone was adopting a Galton/Henry system of identification by

10

Scotland Yard in 1900. Since then fingerprints became one of the most important

features used in forensic prosecutions [22].

Fingerprint identification is also known as dactyloskopy or also hand identification is

the process of comparing two examples of friction ridge skin impression from human

fingers, palm or toes[24].

There are a lot of easy to use and cheap fingerprint scanners. They are based on

different technologies including optic, capacitive, ultrasound, pressure, thermal and

electric field sensors. (figure 2.1)

The technology is widely accepted as very reliable. There is a common belief

(however never proved!) that fingerprints are unique in whole human population.

Finger print is also stable and doesn‟t change with time; it remains same for all the

period of life. That is why fingerprint evidence is even acceptable in a court of law.

Figure 2.1: Finger print [22]

Advantages of Finger Print Recognition

The finger print can be easily taken

Even in twinsvary

Systems Require Less Space

Disadvantages of Finger Print Recognition

Public Perceptions

PrivacyConcerns of Criminal İmplications

11

2.5.2 Face Recognition

Through the whole life of humanity, face was used as the first and main recognition

methods of people. Facial recognition systems analyze facial characteristics. This

system requires a digital camera or a camcorder to develop a facial image of the user

for identification. Facial recognition technology is claimed to be the fastest growing

area in biometric technologies [25].

Face recognition is one of the promising techniques nowadays. The possibility of

covert identification of people unaware of that makes it eligible for – for instance –

terrorist search in crowded places. First face recognition technologies were the so

called geometric based methods.(figure 2.2). They were based on recognition of the

specific elements of human face like nose or eyes and measuring its relative positions

and shape. The methods were insensitive to variations in illumination and viewpoint

[22]. Many approaches have been proposed for face recognition; such as eigen faces

transform. This transform based on creating many images from the source containing

most of the meaningful parts of the image. Other methods include Filcher Linear

Discriminant Analysis or Independent Component Analysis. The most recent is the

3D face recognition approach.

Although the face recognition field has been widely developed recently, the need for

more and more secure and accurate recognition methods is still demanding.

Figure 2.2: Face recognition (2D technology) [24].

12

2.5.3 DNA

DNA is a part of the cell that contains genetic information unique for each person.

DNA typing is a method of biometrics that measures and analyses DNA to distinguish

people with some degree of probability. This method of biometrics is rather popular in

criminalistics. It should be mentioned that DNA is the only method of biometrics that

is not automated and it takes hours to make the DNA test. But this method is

considered to be one of the most reliable methods; and due to it many unsolved

crimes were solved [24]. The only exception for DNA is the case of twins where the

DNA can be the same.(figure 2.3)

Figure 2.3: DNA [24]

2.5.4 Iris

Visual texture of the human iris is determined by the chaotic morphogenetic processes

during embryonic development and is posited to be unique for each person and each

eye [26]. The primary visible characteristic of iris is the trabecular meshwork, that

makes possible to divide the iris in a radial fashion. It is formed in the eighth month of

gestation. Iris is stable and does not change during the whole life [24].

13

Iris recognition is considered to be one of the exact methods of recognition. The

damage of iris features is less possible compared to the finger or palm print due to the

eyelids. One of the main advantages of iris recognition is its stability during the whole

life, while its small size(figure 2.4) and impossibility of taking its photo from a

distance are the most disadvantages of this method. Also people with blindness or

cataract are difficult to be involved in iris recognition process as it is difficult to

extract information from their iris.

Figure 2.4: Iris [26]

2.5.5 Palm Print

Palm-print is often mentioned with such methods as fingerprints and iris recognition.

Palm-print is also distinctive and can easily be captured with low resolution devices

[24]. The inner surface of palms, has rich features including principal lines, wrinkles,

minutiae points, singular points and texture. These features can be used for uniquely

identifying a person. Currently, there are two types of palm-print research, high

resolution approach and low resolution approach.(figure 2.5) High resolution

approach is suitable for forensic applications while low resolution approach is suitable

for commercial applications [21].

Human beings were interested in the palm lines for fortune telling long time ago. In

this century, scientists discovered that the palm lines were associated with some

genetic diseases including Down syndrome, Aarskog syndrome, Cohen syndrome and

fetal alcohol syndrome [21]. The palm of each person consists of principle lines,

wrinkles secondary lines and ridges. Palm also contains such information as texture,

14

indents and marks which are used during the comparison of one palm with another

[24].

Figure 2.5: High and low resolution palm print images [27].

2.5.6 Signature

It is a behavioral biometric, is widely accepted in governmental, legal and commercial

transactions. Each person can have several signatures for different applications.

Nevertheless, a signature cannot uniquely identify a person. Many factors can

influence the consistency of signatures such as emotional and physical conditions.

Furthermore, professional forgers are capable of reproducing signatures to fool

recognition systems [21].

2.5.7 Voice

Voice for humans is claimed to be unique like walking way and sent. It is how ever

taking long time to analyze the voice and identify a person. To identify the person

with the help of voice print, a sample of speech should be taken. This sample is

analyzed. Different multiple measurements are taken and the results are presented in

the form of the algorithm.

15

2.5.8 Ear

Ear recognition has received considerably less attention than many alternative

biometrics, including methods based on face, fingerprint, iris, and gait recognition

[27]. However, recently ear features have been used for many years in the forensic

science for recognition. Ear is stable and unique for each person that can‟t change

during the life [28]. The structure of ear is not completely random. As all other

biometrics, ear has standard features.(figure 2.6) But in the contrary of the human

face, ear doesn‟t have expression changes and make-up effects.

Figure 2.6: Ear [28]

Ear biometrics can be popular for human recognition for many reasons [9]:

1. Ear is found to be very stable. Medical studies have shown that major changes in

the ear shape happen only before the age of 8 years and after that of 70 years. Shape

of the ear is found to be stable for rest of the life.

2. Ear is remarkably consistent and does not change its shape under expressions like

face.

3. Color distribution of the ear is almost uniform.

16

4. Handling background in case of ear is easy as it is very much predictable. An ear

always remains fixed at the middle of the profile face.

5. Ear is unaffected by cosmetics and eye glasses.

6. Ear is a good example of passive biometrics and does not need much cooperation

from the subject. Ear data can be captured even without the knowledge of the subject

from a distance.

7. Ear can be used in a standalone fashion for recognition or it can be integrated with

the face for enhanced recognition.

Researchers have suggested that the shape and appearance of the human ear is unique

to each individual and relatively little change occurs during the lifetime of an adult

[2]. The potential for using the ear‟s appearance as a means of personal identification

was recognized and advocated as early as 1890 by the French criminologist Alphonse

Bertillon. Unlike face recognition, ear is also not affected by factors such as mode or

health. The outer shape of ear remain same and is not affected by age [1].

2.5.9 Other Biometrics

Other biometrics includes gaits, lip prints, brain signals, teeth, retinas, odor,

keystrokes, heights, weights and genders have been proposed. They have different

characteristics and different potential applications.

17

CHAPTER THREE

ARTIFICIAL NEURAL NETWORKS

3.1 Overview

The idea of artificial neural network appeared in the 5th decade of the 20th century. It

was firstly a try of imitating the structure and function of the human brain. However,

the philosophical and psychological perspective of such ideas were originated and

studied by great thinkers like Plato and Aristotle. The first paper on neural network

was published in 1943 by McCuloch and Pitts presenting a simple neuron that

produces 0 or 1[30].

This chapter of the thesis will be concerned by the study of the structure and function

of Artificial Neural Network. It will be especially discussing the back-propagation

learning algorithm of ANNs.

After the first paper proposing the neural network as a processing network, many

works has been proposed successively. In 1949, Hebb proposed a learning process

which has become the starting of the neural network learning algorithms. In 1962,

Rosenblatt was able to prove the convergence of a learning algorithm. that means the

iterative adjustment of the weights of the network until obtaining a desired set of

outputs. The use of single layer neural networks was unable to present an efficient

solution for a lot of types of problems. Although it was believed that multilayer

network can offer better performance for such problems, but there was no efficient

learning algorithm which can insure the convergence of the network.

The revolution in the development of ANNs happened in the middle of the 9th

decade

of the 20th

century. When researchers invented an algorithm for training multilayer

ANNs.

3.2 Introduction of ANNs

An artificial neural network is a system composed of many simple processing

elements operating in parallel to compose a functional network. The function of these

elements is determined by the structure of the network, the strength of the connections

between its elements, and the processing manner performed in the nodes. The

structure of ANNs is inspired by the architectural structure biological nervous system.

18

The biological nervous system uses millions of interconnected parallel elements to

obtain high computing efficiency.

An artificial neural network is a huge parallel processor having natural propensity for

storing experiential knowledge and making it available for use [29]. It resembles the

brain in two aspects:

Knowledge is gained by simple learning process.

The information is stored in the synaptic weights which represent the inter-neural

connection strength.

The neuron is like a MISO (multiple inputs single output) system, its output strength

is determined by the strengths of its different inputs. The signals of all neurons are

summed and compared to a threshold defined by the user to determine if the output

shall be excited.

3.3 Biological Neuron

The brain is composed of billions of neurons interconnected between each other with

billions of interconnections. Each biological neuron is composed of five parts: cell

body, nucleus, dendrites, synapse, and axons.(figure 3.1) Dendrites are connected to

the cell body. They are responsible for receiving a signal from the connection point

called synapse. Synaptic junction can both receive and transmit information. The

received signal is transmitted chemically changing the electrical potential inside the

cell body. If the potential exceed a defined threshold, a pulse of defined strength and

threshold is generated through the axon to the other neurons [30]. The neuron is able

to respond to the total of its inputs aggregated within a short time interval called the

period of latent summation. The membrane can be considered as a shell, which

aggregates the magnitude of the incoming signals over some duration. Specifically,

the neuron generates a pulse response and sends it to its axon only if the conditions

necessary for firing are fulfilled [31].

19

Figure 3.1: Basic biological neuron [31].

3.4 Artificial Neuron And Neural Networks

Artificial neural network (ANN) is a discipline that draws its inspiration from the

incredible problem-solving abilities of nature‟s computing engine, the human brain,

and endeavors to translate these abilities into computing machine which can then be

used to tackle difficult problems in science and engineering. However, all Artificial

neural network paradigms involve a learning phase in which the neural network is

trained with a set of examples of a problem [31]. An artificial neuron is designed to

mimic the basic functions of biological neurons [30]. It computes the sum of the

weights applied to its inputs. This sum is the equivalent of the electrical potential

transmitted through the cell body in the biological neuron. The output is then passed

to an activation function to determine if it is enough to activate the neuron or not. The

equation describing the sum in the ANN is: (figure 3.2)

ij iij

NET w x (3.1)

20

Figure 3.2: Sample structor of ANN

3.5 Adaptive Networks And Systems

3.5.1 Activation Function

After the processing of inputs with associated weights and finding the sum of them,

an activation function is used to determine whether the output is to be activated or not.

Some activation functions are used to determine how much the processed input will

share in constructing the total output of the network. There are many types of

activation functions in artificial neural networks:

3.5.1.1 Linear Activation Function

This type of activation function is known by its function defined by: (figure 3.3)

1 1

( ) 1 1

1 1

x

f x x x

x

(3.2)

21

Figure 3.3: Linear activation function

3.5.1.2 Non-Linear Activation Functions

There are several types of non-linear activation functions; the two most common are

the log-sigmoid transfer function and the tan-sigmoid transfer function. Plots of these

differentiable, non-linear activation functions are illustrated in figure. They are

commonly used in networks trained with back propagation. The networks referred to

in this work are generally back propagation models and they mainly use log-sig and

tan-sig activation functions [31]. The logarithmic sigmoid function is defined by its

formula:

( ) 1/ (1 exp( ))f x x (3.3)

σ =1 though it can be changed which in turn changes the shape of the sigmoid. As σ

tends toward infinity it behaves more and more like a hard-limiter where the slope of

the sigmoid is zero. In this case where the slope is not zero, the output range is

contained between 0 and 1. The tangential sigmoid function is defined by: (figure 3.5)

( ) (1 exp( ) / (1 exp( ))f x x x (3.4)

22

Where it differs from the logarithmic sigmoid in that its output is limited by the -1 and

one as illustrated in figure.(figure 3.4)

Figure 3.4: Logarithmic sigmoid activation function.

Figure 3.5: Tangential sigmoid activation function

-100 -80 -60 -40 -20 0 20 40 60 80 1000

0.2

0.4

0.6

0.8

1

x

f(x)

-100 -80 -60 -40 -20 0 20 40 60 80 100-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1

x

f(x)

23

3.5.1.3 Hard Limit ( Threshold Function)

A threshold (hard-limiter) activation function is either a binary type(figure 3.6) or a

bipolar type(figure 3.7) as shown in Figures.

Figure 3.6: Binary activation function

Figure 3.7: Bipolar activation function

3.6 Learning Methods of ANN

In general, learning is a relatively permanent change in behavior brought about by

experience. Learning in neural networks is a more direct process, and we typically can

capture each learning step in a distinct cause-effect relationship. The knowledge of a

-3 -2 -1 0 1 2 3

0

0.5

1

1.5

2

x

f(x)

f(x)

-3 -2 -1 0 1 2 3-2

-1.5

-1

-0.5

0

0.5

1

1.5

2

x

f(x)

24

neural network is stored in the synapses, which are the weights of the connections

between the neurons [31]. A neural network has to be configured such that the

application of a set of inputs produces the desired outputs. Various methods to set the

strengths of the connections exist. One way is to set the weights explicitly using a

priori knowledge. Another way is to train the neural network by feeding it teaching

patterns and letting it change its weights according to some learning rule [32].

Dependent on the previous context, the learning methods of ANN can be divided into

two categories: supervised and Un-supervised learning methods.

3.6.1 Unsupervised Learning

In this method, a unit is trained to respond to clusters of pattern within the input. In

this paradigm the system is supposed to discover statistically salient features of the

input population. Unlike the supervised learning paradigm there is no a priori set of

categories into which the patterns are to be classified rather the system must develop

its own representation of the input stimuli [32].

3.6.2 Supervised Learning

In this method of learning, the network is trained a prior by providing it with

examples. These examples are exactly the same as those given to a student in school.

the network is given inputs examples with associated outputs. After each train of the

network, a measure of the learning error is applied and the weights of the network are

adjusted based on that measure. The training is stopped whenever the error between

the target and actual output of network is acceptable or a maximum number of trains

is reached. One of the most important supervised learning methods which will be our

subject of study is the back-propagation algorithm. Back propagation algorithm will

be explained and discussed in details in the next few pages. It will be used then for the

rest of our work as it has proved its efficiency as a learning method for ANN

applications.

3.7 Back Propagation Learning Algorithm of ANNs

A back propagation neural network uses a feed-forward topology, supervised

learning, and back propagation learning algorithm. This algorithm was responsible in

large part for the re-emergence of neural networks in the mid 1980s. Back

propagation is a general purpose learning algorithm.(figure 3.8) It is powerful but also

25

expensive in terms of computational requirements for training. A back propagation

network with a single hidden layer of processing elements can model any continuous

function to any degree of accuracy (given enough processing elements in the hidden

layer) [33].

Figure 3.8: Back propagation network

Although many improvements on basic back propagation algorithm have been

proposed and applied, still the basic algorithm is the most widely used and famous

one. That is mainly due to the fact that it is simple and easy to be understood; in

addition to that it works for a wide range of problems [33]. Back propagation can be

divided into three steps:

The input data is presented to the input layer of the ANN, the input non processing

layer passes this data to the hidden layer where the data is processed with weights and

using summing and activation functions. The output of the hidden layers is then

passed to the output layers. The input data of output layer is processed as in hidden

layers. The last output is then compared to the desired targets and error is generated.

26

The generated error is then back propagated through the output and hidden layers; that

is the weights of the output and hidden layers are readjusted using formulas of the

error which guarantee the convergence of the error toward null value.

After the adjustment of the weights, the inputs are passed again to the input, hidden,

and output layers and a new error is calculated in a second iteration and vice versa.

The three mentioned step continue being applied until the error of the training

becomes less than an acceptable set value or a maximum number of iterations is

achieved. The flowchart shown in figure (3.9) explains the Back Propagation

algorithm.

Two major learning parameters are used to control the training process of a back

propagation network. The learn rate is used to specify whether the neural network is

going to make major adjustments after each learning trial or if it is only going to make

minor adjustments. Momentum is used to control possible oscillations in the weights,

which could be caused by alternately signed error signals. While most commercial

back propagation tools provide anywhere from 1 to 10 or more parameters for you to

set, these two will usually produce the most impact on the neural network training

time and performance [33].

27

Start

Reading Images and ANN

Parameters

Preprocessing of the images in MATLAB

Set random hidden and output weights and biases

Construct the input Outputmatrix of

ANN

Training of the network using Back Propagation

Algorithm, calculation of outputs for each pattern

Calculating the MSE

Preprocessing of the images in MATLAB

If MSE<Epsilon

Or

Number of Iterations > max. iterations

Yes

No

Save the last Parameters and weights

Display results and Error

End

Figure 3.9: Block diagram of the training process (flowchart)

28

The back propagation is a method based on error minimization theory and by

searching the least squared error using the gradient descent. Since this method

requires computation of the gradient of the error function at each iteration step, we

must guarantee the continuity and differentiability of the error function at all time.

That implies the use of differentiable continuous function as an activation function.

This condition can be guaranteed by the use of the sigmoid tangent or logarithm

mentioned previously. The logarithmic sigmoid defined by the function (3.5) is

differentiable and its derivative is given by (3.6).

1( )

1 xf x

e

(3.5)

The constant c in this function can be selected arbitrarily and its reciprocal 1/c is

called the temperature parameter in stochastic neural networks [34]. The shape of the

sigmoid logarithmic function with different c values is shown in figure (3.10). Higher

values of c bring the shape to a step function shape and an infinite value of c make the

sigmoid to converge to a step function.

Figure 3.10: Different sigmoid plots for different values of „c‟

Deriving the sigmoid logarithm function gives:

2

( ) ( )(1 ( ))(1 )

x

x

d ef x f x f x

dt e

(3.6)

-100 -80 -60 -40 -20 0 20 40 60 80 1000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

x

f(x)

c=0.05

c=0.1

c=1

c=1.9

c=10

c=20

29

3.7.1 Learning Problem

The neural network represents a chain of function compositions which transform an

input to an output vector (called a pattern). The network is a particular

implementation of a composite function from input to output space, which we call the

network function. The learning problem consists of finding the optimal combination

of weights so that the network function approximates a given function “f” as closely

as possible. However, we are not given the function “f” explicitly but only implicitly

through some examples [34].

Considering a neural network with n inputs and m outputs. It can also consist of any

arbitrary number of hidden units. The weights of the network are real numbers

selected randomly. Given a set of inputs and outputs {(x1, t1 ),(x2, t2 ), . . . , (xp, tp )}

consisting of p ordered pairs of n and m vectors representing the input and output

patterns. When the input xi from the training set is presented to the network, it

produces an output oi generally different from the target ti. Our goal is to have

identical outputs oi and targets ti for all inputs and outputs of the set. In other words,

the aim is to minimize the error function of the network defined by:

2

1

1( )

2

p

i iit o

(3.7)

After minimizing this function for the training set, new unknown input patterns are

presented to the network and we expect it to interpolate. The network must recognize

whether a new input vector is similar to learned patterns and produce a similar output

[34].

The back-propagation algorithm is used to find a local minimum of the error function.

The network is initialized with randomly chosen weights. The gradient of the error

function is computed and used to correct the initial weights. Our task is to compute

this gradient recursively. The network is said to be learned if the error arrives a

minimum acceptable value and all or most of patterns are recognized correctly.

After the end of learning process of network, arbitrary patterns can be presented to the

network in a test process. The results of the network are then threshold using a

threshold chosen based on the application. High threshold values can be used in

30

health, military, or security applications, while smaller thresholds can be accepted for

less important applications.

This chapter has discussed the structure and function of neural networks. The

development of the ANN has be presented through the 20th

century has been

presented. A study of back-propagation algorithm has been presented and discussed.

The application of back propagation learning algorithm for the ear recognition will be

discussed and studied in the next chapter. Samples from 54 persons left and right ear

were taken and presented to the neural network for learning process. Results and

discussions about the used methods and procedures will be detailed.

31

CHAPTER FOUR

EAR RECOGNITION EXPERIMENTAL RESULTS

4.1 Overview

An ear recognition system must deal with two basic problems; the extraction of the

most important features of the ear, and the recognition of these ears or for whom they

belong.

This chapter presents a detailed discussion about the back-propagation based ear

recognition system. The use of the back propagation in the recognition of different

right and left ear images is discussed and presented. 54 different pairs of ears are used

in the identification of 54 persons who are the sample of this thesis. The ears taken

from these sample persons are then treated with different types of noises to provide

different example for the neural network. The sets of real and noisy images are

preprocessed and then presented to the ANN using the back propagation learning

process in the training phase. After that a simple test is applied to check the ability of

network to recognize processed ear photos.

The chapter presents the methodology of the research and the different steps of it;

starting by collecting the data base and ending with the testing phase. The application

of the noise and processing of the images is an important phase in the recognition

procedure. It is discussed in this chapter with presentation of different processed

features and their effect.

The processing of the images data using MATLAB including adding the noise to the

pictures in addition to resizing and changing the type of images to reduce processing

cost. The pre-processing phase is a very important phase for a successful recognition

system. The choice of image sizes in addition to the methodology of working affect

directly the results of the program in addition to the time of working.

4.2 Database Collection

The first step in this work was the data base collection. All the images where collected

from the students of the computer engineering department in Near East University.

32

The photos of ears of 54 students were taken using a 13 megapixel Samsung camera.

Photos included the right and the left ears images for each person. A distance of 10

centimeters was arbitrary chosen between the camera‟s lens and the ear. An angle of

90 degrees was used to take the photos.(figure 4.1) The photos were then arranged

and processed using Photoshop program. The photos were prepared such that just the

features of the ears appear in the picture. Any empty spaces or other features like hair

where manually removed from the pictures. The preprocessing phase includes the

changing the images to gray scale images, resizing images, normalization of images

data, and adding noise to images.(figure 4.2)

Figure 4.1: Ear photo collection data

33

Read the RGB image

Convert RGB image to gray

scale image

Reduce the size of image

to 50*50

Add different types of Noise

to the images

Normalizing the pixels

values of the image

Convert the image matrix to

vector of size 50*50

Building the input training and

test matrix containing all images

vecors

Send the input matrix to the network

training and test processing

Figure 4.2 : Block diagram of the preprocessing phase of the training and test images.

4.3 General Experiment

Many experiments were carried out on the set of image until arriving suitable

parameters of the neural network in the training process. In the training process of the

neural network 8 sets of normal and noisy images were used. The first set represented

the normal images of the ear. The other sets were noised using MATLAB program.

Different types of noises were used in the training like Gaussian and salt and pepper

34

noise. The MATLAB code used for adding different types of noise in addition to the

noise rates are presented in Appendix A,

The Noises added here in the training process were:

- Gaussian white noise of mean 0.02 and variance 0.0018 and 0.005 to the

normalized image.

- Speckle with multiplicative noise using the equation J = I + n*I, where n is

uniformly distributed random noise with mean 0 and variance 0.04 and 0.08.

- Salt and pepper noise with density 0.05 and 0.01.

A sample of the original and noisy images is shown in figure 4.3 and figure 4.4:

Figure 4.3: Sample of the images used in the training (first person left).

Figure 4.4: Sample of the images of training (first person right).

35

The images of the figure 4.3 and 4.4 represent the training sample of the first person.

These images were preprocessed before being fed to the network. The preprocessing

of the images passes by different phases. The first phase after reading the image in

RGB scale is to change it into gray scale image. The gray scale image represents each

pixel of the image by an unsigned eight bit integer (0-255). This number is the

concentration of white color in the pixel. The pixel is black if its value is zero,

increasing the value increase the white concentration until it maximum in the white

color with the value of 255. In the rgb scale images each pixel is presented using 3

different values; each one of these three values presents the concentration of the three

base colours, red, green, and blue. Using the gray scale image reduces the image size

by two thirds of its original rgb scale size. That operation reduces the calculation

efforts carried by the program with aboutly no effect on the accuracy of the program.

After the process of changing image scale, the size of the image has to be reduced to

an acceptable size to make easier the operations on smaller images in addition to

reduce the processing cost of the program. The size of images was chosen to be

50*50. Figure 4.5 shows the original rgb image and gray scale image in addition to

the resized image.

Figure 4.5: Original RGB image(size(3100*2100)) and Gray scale

image(size(3100*2100)).

RGB image GRAY scale image

36

Figure 4.6: Resized gray scale image (size(50*50)).

The resized image is after that normalized in order to limit the inputs of the ANN by 0

and 1. This operation reduces the processing costs and learning time. The processed

images must be arranged and fed to the neural network such that the one ear

information is processed in each iteration. In the next iteration, an image of a different

ear should be presented to the network until the last image. After finishing all images,

the operation must be repeated from the beginning until the network learns. In order

to make easier the mentioned operation, a small routine was written using MATLAB

to arrange each picture‟s pixels in a vector. The vectors are then arranged in a big

matrix which will be fed column by column to the network. The desired output of the

network was chosen arbitrary and also arranged in an output matrix which was

presented to the network.

4.4 Training of the network

As mentioned in the previous chapters, back-propagation learning algorithm was used

in the training process of the network. After feeding the pictures to the network, three

different experiments were carried out with different network parameters in order to

evaluate the performance of our procedure. In each experiment different parameters of

the network were used. The test was performed with different pictures (noise level)

for each experiment.

37

Experiment 1:

The back-propagation process was started with the next parameters:

Table 4.1: Parameters of the network in first experiment.

Number of input neurons 2500

Experimental number of hidden neurons 200, 240

Number of output neurons 54

Experimental initial learning rate 0.09

Experimental momentum factor 0.3

Minimum error 0.0005 achieved

Number of iterations 4037

Maximum iterations 20000

Training time 20 min 55 sec

Testing time 0.110756 sec

Figure 4.7 shows the mean squared error curve obtained in the training process. In the

hidden layers two activation functions were used, the first is logsig activation

function, while the second was tansig function; a logsig function was also used in the

output layer.

Figure 4.7: Curve of MSE in the training.

0 500 1000 1500 2000 2500 3000 3500 4000

10-3

10-2

10-1

Epocs

MS

E

MSE Curve

38

Figure 4.8: Curve of the learning Rate.

We can notice from the training results of this experiment that most of the samples of

ears (including left and right ears) were recognized successfully with minimum of

0.74, the threshold was chosen to be th=0.8 in this experiment. The curve of the error

is showing that the error was decreasing slowly until arriving an acceptable value. The

value of the learning rate during the learning process was variable, but still less than

10 for most of the time as shown in the figure 4.8 After the end of training, a one

iteration test was established to check the ability of trained network to recognize other

images than the training set. The test was applied to two different sets of images A

and B with different noise parameters. Each set contains 4 images for the left ear and

4 images for the right ear of each person. Table 4.4 in the appendix shows the training

results while table 4.5 and 4.6 shows the test results of A and B set. From the tables

4.4, 4.5, and 4.6 we can notice that the recognition of the training set was successful

with recognition rate of 97.9% for a threshold of 0.8. Meaning that 846 images out of

864 were recognized correctly with a rate of more than 0.8. The test of the first set of

images was successful also with recognition of 402 images out of 432. The

recognition rate was 93% with a threshold of 0.5. Concerning the second set B, 346

images out of 432 were recognized correctly with a rate of 80% which seems to be

perfect under the high noise conditions. The time of training in this experiment with

about 20 minutes and 55 seconds in 4037 iterations. The testing time was 0.11 second.

0 500 1000 1500 2000 2500 3000 3500 4000 45000

20

40

60

80

100

120

140

160

Epoch

Learn

ing r

ate

Curve of learning rate

39

Experiment 2:

In this experiment, the parameters of the network have been changed and new training

and test processes have been applied. Results of test and training were presented and

tabulated. The training parameters are given in the next table:

Table 4.2: Parameters of the network in second experiment.

Number of input neurons 2500

Experimental number of hidden neurons 100, 150

Number of output neurons 54

Experimental initial learning rate 0.09

Experimental momentum factor 0.2

Minimum error 0.00005

Number of iterations 17244

Maximum iterations 20000

Training time 103min 10sec

Testing time 0.085397

The curve of the error is shown in figure 4.9, where we can notice that it was

decreasing with good rate and reached a good value. A minimum error of 0.00005

was reached after 103 minutes and 10 seconds with total of 17244 iterations. Also the

achieved MSE was better.

Figure 4.9: MSE curve in experiment 2.

100

101

102

103

104

105

0

0.05

0.1

0.15

0.2

0.25

Epocs

MS

E

40

Table 4.7 in the appendix shows that all the training images were recognized

successfully in this experiment with a rate of 100% under a threshold of 0,8. The

training results show that the network was perfectly trained with high efficiency. For

the test purpose, the sets of images A and B were passed through the network and the

results were checked. Table 4.8 shows that 97.6% of the images of set A were

recognized with threshold of 0.5, 422 images out of 432 were perfectly recognized.

While for the set B results we can notice that just 373 images out of 432 images were

recognized using the same network due to the high noise effects added. The

recognition rate was 86.3% which is considered excellent. The results are shown in

Table 4.9. The difference in efficiency between the recognition results of sets A and B

is due mainly to the amounts of noise added to both sets. The images of set B were

noisier and less clear than the set A. a recognition rate of 86.3% can be considered

high under such conditions of noise.

Experiment 3:

In this experiment, different parameters of the neural network were applied for

training process. After the end of training, a one iteration training efficiency check

and test process were applied. The results of training as like as the test were tabulated

and discussed. The following parameters were applied for the training of the network:

Table 4.3: Parameters of the network in the third experiment.

Number of input neurons 2500

Experimental number of hidden neurons 240, 200

Number of output neurons 54

Experimental initial learning rate 0.09

Momentum factor 0.3

Minimum error 0.0005

Number of iterations 4654

Maximum iterations 20000

Training time 25min 30sec

Testing time 0.088273 sec

41

As shown in the table, the network has achieved an error of 0.0005 in 25 minutes and

30 seconds with 4654 iteration. The training time has been reduced to 25 minutes and 30

seconds. The time of one iteration test was 0.0882 second. The results of this experiment have

shown good recognition rate and high performance. As shown in tables 4.10, 4.11, and

4.12 we can notice that the training recognition was 99.3% with 658 recognized

images out of 864. The threshold was chosen to be 0.8 while in the test a value of 0.5

was used. The set A test gave a recognition rate of 94.4% with 408/432 images

recognized successfully. In the set B, 358 images out of 432 were recognized with a

rate of 82.8%. We can notice also that the recognition of set A was better than Set B.

The results obtained from the different experiments prove the efficiency of the

artificial neural networks in human identification using ear patterns. The next table

summarizes the obtained results in our experiment with all recognition rates:

Table 4.4: Recognition rates of the experiments carried out in this work.

Training recognition

rate. Thr = 0.8

Test recognition rate (Set

A) Thr =0.5

Test recognition rate

(Set B) Thr =0.5

Experiment 1 97.9% (846/864

recognized images)

93% (402/432

recognized images)

80% (346 recognized

images)

Experiment 2 100% 97.6% % (422/432

recognized images)

86.3% % (373/432

recognized images)

Experiment 3 99.3% (658/864

recognized images)

94.4% % (408/432

recognized images)

82.8% % (358/432

recognized images)

As Shown in the table 4.4 showing the training and test results, we can notice that our

training results were successful with rates of 99% remembering that our threshold for

training was 0.8 which is considered high. Compared to other recognition works, the

test results of the set A were high and shown a very good results. The results of the

Set B test were in the order of 85% which is slightly less than the results mentioned in

literature; this can be explained by the high noise levels added to this set‟s images.

42

CONCLUSIONS

There is an increasing need for more secure and easy identification methods. The use

of passwords and security codes is becoming more and more difficult with the

development of information technology. Over more, the explosion in the internet

services, banks services, credit cards, debit cards involve the need for remembering a

huge number of pass phrases which – in addition to be very hard task – can be lost

easily and cost a lot to reinitialize. The need also for more secure systems in airports

and police services in order to fight the crime has increased. All these factors have

pushed the researchers toward finding more natural, easy, and highly secure

identification methods. One of the most important methods of identification which

became widely used is the biometrics. Although the use of ear as unique and stable

feature of the human being has been introduced into biometrics since many years, it is

still need for more attention of the researchers due to its simplicity and good

performance.

This work has been carried out in the goal of giving more attention for the use of ear

as an identification method. Combining the use of ear as a biometric feature with the

use of artificial neural network can be considered as a promising idea for the near

future as the neural networks is showing very fast development in the last years.

Artificial neural networks have proved its ability for solving many non-linear

problems efficiently with a minimum effort.

From the experiments carried out in this thesis and the results obtained we conclude

that the use of the neural network for ear recognition and persons Identification was

successful. The application of different noise on the ear‟s images has led to reduction

of recognition rate. Many experiments were carried out with different network

parameters and noise levels. The first experiment has given a training recognition rate

of 97.9%. The test recognition rate of 93% for the set A was achieved. Increasing the

level of noise has affected slightly the recognition capability of the neural network;

that is, in set B the recognition rate has been reduced to 80%. The training time in the

first experiment was about 20 minutes and can be considered fast. In a second

experiment the parameters of the network has been changed. The maximum error has

been changed to 0.1 of that used in experiment one. The recognition rate of 100% in

the training and 97.6% for the test of set A was achieved. The recognition rate of set

43

B was 86.3%. The reduction of maximum error has increased the accuracy of the

network recognition, but the time of training increased to more than100 minutes. This

is relatively long time if compared to the training time of first experiment. The third

experiment has given a recognition rate of 99.3% in the training, with 94.4% and

82.8% recognition rate in the test of set A and B respectively.

Using the network for the identification of both right and left ear images at once has

shown a good level of success with high performance. After finishing the work we

can conclude that the use of ANN in the ear recognition for person identification can

be achieved with high accuracy and can give good levels of security. The neural

networks can be considered as a very good method for recognition and biometrics

identification.

The results obtained in this work proved that the use of ANN for persons‟

identification using ear recognition is a promising idea. The results obtained were

very successful in term of identification rate and time efficiency. The work carried out

in this thesis directs toward making more effort in the subject and paying more

attention for the ANN based ear recognition process. It opens the doors for more

intense and large work in the science of biometrics. As a future work, this thesis

proposes the use of multi-modal identification based on artificial neural network and a

comparative work between the ANN based ear recognition and the other methods

such as wavelet transform based ear recognition.

44

REFERENCES

[1] Mohamed Ibrahim Saleh, “ Usıng Ears For Human Identıfıcatıon”, Master Thesis,

Blacksburg, Virginia May 2007.

[2] Ping Yan, B.S., M.S. “ Ear Biometrics in Human Identification ”, Graduate

Program in Department of Computer Science and Engineering Notre Dame, İndiana

June 2006.

[3] Alastair H. Cummings, Mark S. Nixon, John N. Carter “Novel Ray Analogy for

Enrolment of Ear Biometrics”, 2010.

[4] Vankadara Nvd Manohar Gaurav Uday Chaudharı Bıswajıt Mohanty,

“Functıon Approxımatıon Usıng Back Propagatıon Algorıthm In Artıfıcıal Neural

Networks”, A Thesıs Submıtted In Partıal Fulfıllment of The Requırements For

The Degree of Bachelor of Technology In Electrical Engineering 2007.

[5] Hazem M. El-Bakry Faculty of Computer Science & Information Systems,

Mansoura University, Egypt&Nikos Mastorakis Technical University of Sofia,

BULGARIA “Ear Recognition by using Neural Networks”, ISSN: 1790-2769.

[6] Hui Chen, Bir Bhanu “Human Ear Recognition in 3D”, IEEE Transactıons On

Pattern analysis, Vol.29, No.4, April 2007.

[7] Md. Mahbubur Rahman, Md. Rashedul Islam, Nazmul Islam Bhuiyan, Bulbul

Ahmed, Md. Aminul Islam. “Person Identification Using Ear Biometrics”, Computer

Science and Engineering Discipline, Khulna University, Khulna-9208, Bangladesh

August 2007.

[8] Sukhdeep Singh, Dr. Sunil Kumar Singla(Assit.Prof.) Electrical & Instrumentation

Engg Dept. Thapar University ,Patiala, India. “ Review on Biometrics and Ear

Recognition Techniques”, International journal of advanced research in computer

science and software engineering. Volüme 3, ıssume 6, june 2013.

[9] Surya Prakash, “Human Recognition using 2D and 3D Ear Images”, PhD Thesis

Indıan Instıtute of Technology Kanpur, Indıa February 2013.

[10] Dasari Naga Shailaja, “A Simple Geometric Approach For Ear Recognition”,

Master Thesis, İndian Instıtute of Technology, Kanpur June 2006.

[11] Mina Ibrahim Samaan Ibrahim, “Wavelet Based Approaches For Detection and

Recognition İn Ear Biometrics”, PhD Thesis, Unıversıty of Southampton June 2012.

45

[12] Mark Burge, Wilhelm Burger, “Ear Biometrics in Computer Vision ” IEEE,2000.

[13] Amioy Kumar, Madasu Hanmandlu, Mohit Kuldeep, And H.M. Gupta,

“Automatic Ear Detection For Online Biometric Applications ”, IEEE 2011.

[14] Ruma Purkait “Ear Biometric: An Aid To Personal Identification”,

Anthropologist Special Volume No. 3: 215-218 (2007).

[15] Barnabas Victor, Kevin Bowyer, And Sudeep Sarkar, “An Evaluatıon Of Face

And Ear Bıometrıcs ”, IEEE 2000.

[16] Ping Yan, Kevinw. Bowyer, “Empirical Evaluation of Advanced Ear Biometrics

”, Proceedings of the Computer Society Conference On Computer Vision And

Pattern Recognition IEEE 2005.

[17] Michał Chora´S, “Image Feature Extraction Methods For Ear Biometrics - A

Survey ”, 6th International Conference On Computer Information Systems And

Industrial Management Applications (CISIM'07) IEEE 2007.

[18] Ping Yan And Kevin W. Bowyer, Fellow, Ieee, “Biometric Recognition Using

3D Ear Shape ”, IEEE Transactıons On Pattern Analysıs And Machıne Intellıgence,

Vol. 29, No. 8, August 2007.

[19] Haijun Zhang, Zengxi Huang And Yibo Li, “An Overview of Multi-Modal

Biometrics Based On Face And Ear ”, Proceedings of The Ieee International

Conference On Automation And Logistics Shenyang, China August 2009.

[20] Daishi Watabe, Hideyasu Sai, Katsuhiro Sakai And Osamu Nakamura, “Ear

Bıometrıcs Usıng Jet Space Sımılarıty ”, Saitama Institute Of Technology, Kogakuin

University, IEEE 2008.

[21] Adams Wai Kin KONG, “Palmprint Identification Based on Generalization of

IrisCode”, A thesis presented to the University of Waterloo in fulfillment of the thesis

requirement for the degree of Doctor of Philosophy in Electrical and Computer

Engineering Waterloo, Ontario, Canada, 2007.

[22] Doctoral thesis Paweł Kasprowski, ” Human identification using eye

movements”, Silesian University of Technology Faculty of Automatic Control,

Electronics and Computer Science Institute of Computer Science, Gliwice 2004.

[23] V.Bhawanı Radhıka, ” Bıometrıc Identıfıcatıon Systems: Feature Level

Clusterıng of Large Bıometrıc Data And Dwt Based Hash Coded Ear Bıometrıc

46

System”, Thesis Submitted n Partial Fulfillment of The Requirements For The Award

of the Degree. Bachelor of Technology (Computer Science And Engg.), May 2009.

[24] Aleksandra Babich,” Biometric Authentication. Types of Biometric

İdentifiers”,Bachelor‟s Thesis Degree Programme in Business Information

Technology 2012.

[25] Cassandra M. Carrillo, Thesis” Contınuous Bıometrıc Authentıcatıon For

Authorızed Aırcraft Personnel: A Proposed Desıgn”, Naval Postgraduate School

Monterey, California June 2003.

[26] Http://Www.Ahmetkakici.Com/Genel/Biyometrik-Tanima-Sistemleri/

[27] Mohamed Ibrahim Saleh,” Usıng Ears For Human Identıfıcatıon”, Thesis

Submitted To The Faculty of Virginia Polytechnic Institute And State University In

Partial Fulfillment of The Requirements For The Degree of Master of Science In

Computer Engineering A. Lynn Abbott, Chairman Chris L. Wyatt Patrick Schaumont

Blacksburg, Virginia, May 2007.

[28] Dasari Naga Shailaja,”A Simple Geometric Approach For Ear Recognition”,A

Thesis Submitted İn Partial Fulfillment of The Requirements For The Degree of

Master of Technology, June 2006.

[29] Bilal F. Alnamiawri, “Sign Language And Gesture Recognition Systemusing

Neural Networks”, A Thesis Submitted To The Graduade School of Applied Sciences

Of NEU 2012.

[30]Nancy Y. Xiao,”Using The Modified Back Probagation Algorithm to Perform Automated

Downlink Analysis”, Master of Engineering Electrical& Computer Science June 1996.

[31] Kiran Kumar Kaki,” Parallelized Backpropagation Neural Network Algorithm

using Distributed System”, Thesis submitted in partial fulfillment of the requirements

for the award of degree of Master of Engineering in Computer Science and

Engineering, January 2009.

[32] Ben Krose Patrick Van Der Smagt,Book ”An Introduction To Neural Network”,

The University of Amsterdam 1996.

[33] Charu Gupta, “Implementatıon of Back Propagatıon Algorıthm (of Neural

Networks) In VHDL”, Thesis Report Submitted Towards The Partial Fulfillment of

Requirements For The Award of The Degree of Master of Engineering (Electronics &

Communication), June 2006.

[34] Chin-Chuan Han, Hsu-Liang Cheng, Chih-Lung Lin, Kuo-Chin Fan,” Personal

Authentication Using Palm-Print Features ”,Pattern Recognition 36 (2003) 371 – 381.

47

APPENDICES

Appendix A

Sample Images of Training Data base

48

49

In the training process, the noise rates used were:

imnoise(cc,'gaussian',0.02,0.0018);

imnoise(cc,'gaussian',0.02,0.005);

imnoise(cc,'speckle',0.04);

imnoise(cc,'speckle',0.08);

imnoise(cc,'salt & pepper',0.05);

imnoise(cc,'salt & pepper',0.01);

imnoise(cc,'poisson');

The Noises added here in the training process were:

- Gaussian white noise of mean 0.02 and variance 0.0018 and 0.005 to the

normalized image.

- Speckle with multiplicative noise using the equation J = I + n*I, where n is

uniformly distributed random noise with mean 0 and variance 0.04 and 0.08.

- Salt and pepper noise with density 0.05 and 0.01.

Appendix B

cc=imnoise(cc,'gaussian',0.02,0.0018);

cc= imnoise(cc,'salt & pepper',0.01);

cc=imnoise(cc,'speckle',0.04);

cc=imnoise(cc,'poisson');

cc= imnoise(cc,'salt & pepper',0.05);

cc=imnoise(cc,'gaussian',0.02,0.005);

cc=imnoise(cc,'speckle',0.08);

50

Noise A

switch k,

case 1,

cc=imnoise(cc,'gaussian',0.03,0.0028);

case 2,

cc= imnoise(cc,'salt & pepper',0.15);

case 3,

cc=imnoise(cc,'speckle',0.12);

case 4,

cc=imnoise(cc,'poisson');

end

- Gaussian white noise of mean 0.03 and variance 0.0028 and 0.005 to the

normalized image.

- Speckle with multiplicative noise using the equation J = I + n*I, where n is

uniformly distributed random noise with mean 0 and variance 0.12.

- Salt and pepper noise with density 0.15.

Noise B

switch k,

case 1,

cc=imnoise(cc,'gaussian',0.03,0.02);

case 2,

cc= imnoise(cc,'salt & pepper',0.25);

case 3,

cc=imnoise(cc,'speckle',0.22);

case 4,

cc=imnoise(cc,'poisson');

end

- Gaussian white noise of mean 0.03 and variance 0.02 to the normalized image.

- Speckle with multiplicative noise using the equation J = I + n*I, where n is

uniformly distributed random noise with mean 0 and variance 0.22.

- Salt and pepper noise with density 0.25.

It is clear that the noise levels of set B are very high compared to those of set A.

51

Appendex C

Table 4.5: Training results.

(%) E1R E2R E3R E4R E5R E6R E7R E8R E1L E2L E3L E4L E5L E6L E7L E8L

P1 94 88 93 92 88 94 90 92 93 92 93 92 93 90 95 92

P2 90 89 91 90 92 93 87 87 91 90 88 92 91 91 91 95

P3 92 94 93 84 90 82 92 85 89 86 90 83 92 89 95 94

P4 93 92 91 89 93 90 92 96 91 92 91 91 91 85 90 92

P5 84 87 84 85 81 93 86 93 89 90 90 88 88 88 91 97

P6 86 79 85 89 87 77 80 80 90 91 92 84 93 85 85 85

P7 93 95 93 93 92 94 92 92 87 86 87 91 87 88 88 92

P8 83 92 87 84 86 81 82 83 86 87 79 83 84 82 91 81

P9 90 91 90 91 87 88 87 86 83 93 84 87 88 82 92 85

P10 91 91 89 89 87 87 95 92 90 94 93 89 86 79 92 93

P11 91 90 92 91 90 94 82 84 92 91 90 94 91 86 92 91

P12 89 92 86 81 88 80 96 93 87 89 86 90 85 78 92 86

P13 82 82 79 78 74 87 89 81 81 74 78 92 84 87 75 87

P14 93 94 92 92 93 92 92 89 86 89 87 93 91 83 91 90

P15 87 85 89 90 89 83 91 93 83 86 86 87 78 85 94 85

P16 92 84 93 84 86 95 94 95 93 80 91 82 94 89 95 80

P17 88 92 90 85 90 89 92 94 94 93 92 86 92 90 91 85

P18 86 86 88 87 90 92 89 91 89 91 85 90 87 86 88 93

P19 89 84 87 90 88 81 85 79 90 89 89 79 92 90 90 95

P20 93 95 94 90 93 95 96 93 89 92 90 92 89 89 92 89

P21 93 95 93 93 92 94 93 92 94 93 94 93 93 84 96 92

P22 93 91 93 93 89 96 87 82 91 92 86 94 90 87 92 93

P23 91 92 90 95 89 93 92 94 92 93 88 95 92 93 92 91

P24 91 84 90 87 88 87 86 95 88 87 89 96 91 83 87 91

P25 88 87 85 90 86 89 80 91 89 88 91 85 88 84 92 82

P26 86 84 85 89 87 95 89 82 85 86 84 75 90 88 78 74

P27 91 91 89 91 90 87 91 82 91 92 93 93 93 91 94 90

P28 88 84 91 80 88 77 87 85 91 93 91 92 89 84 91 93

P29 91 84 93 89 91 86 88 81 87 85 85 90 86 86 89 84

52

P30 91 86 90 89 90 82 85 92 85 80 91 93 88 79 84 85

P31 93 93 93 95 93 91 93 94 92 92 93 94 90 89 90 96

P32 92 93 91 90 94 85 94 90 92 96 90 93 94 83 94 91

P33 92 84 90 87 88 93 88 83 92 87 91 93 93 84 87 89

P34 87 88 86 87 83 90 85 83 92 90 93 87 90 91 82 89

P35 91 89 89 90 91 85 93 93 87 89 87 90 88 90 90 81

P36 91 88 91 92 91 90 89 92 91 90 91 89 90 81 92 92

P37 91 87 90 91 92 93 88 87 92 90 91 91 93 95 88 89

P38 87 85 85 83 88 85 86 91 89 87 86 92 92 79 92 94

P39 89 91 87 83 91 88 91 83 92 94 89 86 91 92 91 89

P40 90 82 84 86 90 85 86 91 90 85 86 91 90 81 90 89

P41 88 83 90 92 88 87 90 91 90 86 91 93 88 91 77 91

P42 87 86 90 83 84 92 94 88 86 88 87 91 86 91 89 90

P43 92 93 90 93 93 85 92 91 93 91 91 93 92 90 93 90

P44 92 92 89 92 93 91 93 89 91 91 91 91 91 89 91 87

P45 91 92 91 95 90 89 94 95 89 88 90 90 84 88 90 81

P46 95 94 95 90 95 89 95 92 94 92 93 91 94 92 93 87

P47 93 93 91 93 93 84 94 91 92 94 89 94 91 85 91 87

P48 96 95 94 94 95 92 95 91 93 91 94 95 95 92 92 95

P49 92 89 92 88 89 88 83 90 92 92 90 89 90 88 88 91

P50 92 84 89 92 88 91 87 89 92 84 94 90 87 91 87 88

P51 90 90 90 91 89 94 87 83 92 88 94 90 91 94 91 93

P52 87 90 85 90 82 87 92 96 88 88 87 81 85 88 87 89

P53 94 96 92 93 93 94 94 89 92 93 91 95 92 88 94 92

P54 90 89 91 92 91 80 85 82 91 89 89 89 87 86 92 93

Table 4.6: Test results: Noise A (30/432 not recognized).

% E1 E2 E3 E4

E1 E2 E3 E4

E1 E2 E3 E4

E1 E2 E3 E4

P1 80 68 88 89 P28 90 84 93 94 P1 83 83 89 87 P28 83 84 86 91

P2 86 91 85 92 P29 89 84 96 89 P2 77 47 82 90 P29 75 46 82 83

53

P3 90 41 69 92 P30 89 75 64 89 P3 78 70 79 90 P30 80 41 72 83

P4 91 91 96 93 P31 91 40 90 92 P4 94 83 90 93 P31 89 93 78 90

P5 86 84 91 82 P32 86 65 97 83 P5 96 46 73 92 P32 96 85 88 92

P6 86 37 75 73 P33 91 41 85 89 P6 85 52 87 90 P33 86 94 93 91

P7 93 82 96 92 P34 83 76 65 88 P7 82 37 93 86 P34 85 73 80 91

P8 82 34 96 87 P35 81 75 52 85 P8 95 80 85 89 P35 88 30 82 87

P9 90 24 66 87 P36 90 24 58 81 P9 81 77 69 87 P36 85 79 86 92

P10 94 62 89 89 P37 94 92 60 92 P10 88 80 89 92 P37 89 92 91 93

P11 85 71 59 86 P38 91 82 90 90 P11 91 33 46 82 P38 91 35 93 91

P12 96 55 55 86 P39 93 53 89 75 P12 90 81 80 88 P39 92 46 77 90

P13 70 54 85 73 P40 82 18 40 82 P13 78 80 94 89 P40 86 26 93 92

P14 94 87 80 94 P41 90 52 88 86 P14 83 61 71 86 P41 88 59 91 89

P15 91 51 77 84 P42 90 65 77 84 P15 88 84 93 88 P42 88 40 90 88

P16 87 50 88 94 P43 94 86 54 88 P16 93 78 88 94 P43 86 74 62 93

P17 89 58 87 93 P44 89 70 78 93 P17 92 49 70 92 P44 89 68 82 90

P18 90 61 91 82 P45 83 71 92 91 P18 93 55 89 93 P45 88 52 72 87

P19 87 62 72 92 P46 86 79 82 87 P19 93 76 84 93 P46 92 82 67 93

P20 96 84 77 95 P47 92 80 71 86 P20 95 75 91 94 P47 92 67 86 91

P21 95 87 86 92 P48 93 53 87 93 P21 94 79 88 96 P48 94 76 93 95

P22 91 49 81 92 P49 86 86 72 90 P22 90 39 53 90 P49 89 80 55 90

P23 92 84 72 92 P50 94 92 94 92 P23 81 90 33 92 P50 76 88 49 86

P24 77 81 81 93 P51 84 48 86 83 P24 92 83 75 90 P51 86 57 78 88

P25 87 34 69 90 P52 82 72 97 85 P25 90 54 75 85 P52 92 74 82 86

P26 76 89 89 86 P53 90 47 74 83 P26 96 77 88 95 P53 94 63 82 90

P27 90 56 85 92 P54 93 77 81 91 P27 86 38 76 87 P54 91 92 79 88

Table 4.7: Test Results, Noise B (86/432 not recognized)

% E1 E2 E3 E4 E1 E2 E3 E4 E1 E2 E3 E4 E1 E2 E3 E4

P1 67 25 73 91 P28 89 43 94 92 P1 87 79 54 87 P28 80 88 76 89

P2 84 90 89 85 P29 91 97 96 89 P2 80 30 75 90 P29 78 4 64 80

54

P3 90 11 54 91 P30 92 56 79 88 P3 58 20 74 90 P30 58 5 77 85

P4 90 20 91 91 P31 88 62 93 92 P4 89 85 93 91 P31 86 84 92 93

P5 74 53 87 77 P32 89 66 86 90 P5 91 40 69 88 P32 90 8 89 92

P6 86 49 79 78 P33 94 46 29 90 P6 84 12 92 92 P33 77 18 92 91

P7 95 89 55 90 P34 92 68 26 85 P7 93 42 66 88 P34 91 50 94 94

P8 95 15 30 83 P35 76 31 25 79 P8 92 10 91 89 P35 80 15 96 89

P9 90 12 21 86 P36 91 16 70 86 P9 52 43 92 88 P36 73 30 87 91

P10 93 15 80 93 P37 95 47 40 90 P10 74 73 79 92 P37 87 41 94 93

P11 68 26 55 87 P38 93 60 72 89 P11 85 20 42 87 P38 81 13 81 88

P12 95 75 43 88 P39 95 6 66 86 P12 85 6 54 85 P39 77 41 76 89

P13 71 21 64 80 P40 59 43 95 89 P13 65 2 95 91 P40 62 61 71 90

P14 96 89 84 95 P41 87 9 70 84 P14 71 10 84 86 P41 28 12 94 88

P15 92 78 92 84 P42 91 64 60 84 P15 90 29 84 88 P42 89 21 91 87

P16 74 39 88 93 P43 84 41 74 91 P16 87 49 89 93 P43 81 45 82 93

P17 72 3 60 87 P44 93 34 66 95 P17 84 27 69 91 P44 89 8 40 89

P18 94 40 57 86 P45 84 54 78 86 P18 74 41 88 92 P45 62 30 60 90

P19 73 40 82 86 P46 88 58 58 90 P19 92 75 41 95 P46 87 61 67 91

P20 96 70 67 93 P47 81 9 44 85 P20 91 34 67 93 P47 92 8 80 91

P21 86 59 53 93 P48 88 48 83 93 P21 91 74 83 91 P48 80 41 90 94

P22 93 51 82 93 P49 74 37 90 88 P22 86 19 33 86 P49 61 41 77 87

P23 92 82 79 90 P50 92 73 84 92 P23 64 76 87 91 P50 88 83 77 92

P24 79 13 58 81 P51 50 57 78 91 P24 96 42 83 92 P51 88 47 53 90

P25 89 83 93 92 P52 88 10 85 89 P25 94 20 67 86 P52 94 75 74 83

P26 63 81 78 87 P53 34 20 47 91 P26 97 57 88 95 P53 89 16 53 92

P27 92 44 63 91 P54 86 72 75 92 P27 87 35 33 87 P54 94 42 75 87

Training results in experiment 2

Table 4.8: Training results in experiment 2

E1R E2R E3R E4R E5R E6R E7R E8R E1L E2L E3L E4L E5L E6L E7L E8L

55

P1 98 94 98 96 96 98 96 98 97 97 96 98 97 96 98 98

P2 97 96 98 97 98 96 95 95 96 97 94 97 96 97 97 97

P3 99 97 98 94 99 92 97 95 96 94 98 93 98 96 96 98

P4 97 97 97 98 97 95 96 99 97 98 97 96 97 95 98 97

P5 96 94 95 95 95 98 94 98 98 97 98 96 97 98 95 97

P6 97 97 98 94 98 91 95 95 97 98 97 94 98 94 95 97

P7 98 98 98 98 98 97 98 95 96 96 96 97 97 98 94 98

P8 96 95 96 96 93 96 95 95 95 96 94 97 93 94 97 95

P9 96 97 97 95 96 97 96 96 93 98 95 95 95 94 97 97

P10 97 97 94 97 96 95 97 98 97 97 98 94 95 92 97 98

P11 97 96 97 97 97 94 97 96 97 97 97 98 96 95 97 98

P12 97 95 96 94 98 96 97 97 95 96 95 98 95 93 97 97

P13 96 96 94 93 93 97 96 96 94 95 96 98 96 95 92 95

P14 98 97 97 98 98 98 98 95 97 97 97 96 97 96 95 98

P15 97 95 96 97 97 93 98 95 95 94 95 97 95 94 98 94

P16 97 93 98 97 96 97 98 96 98 93 96 94 98 96 98 94

P17 96 98 97 96 97 98 95 96 98 98 96 96 98 98 96 98

P18 95 93 97 98 97 98 96 97 98 97 97 97 96 96 93 97

P19 96 97 95 95 96 94 97 94 97 98 96 95 98 97 97 97

P20 97 99 97 96 97 98 98 98 96 98 97 98 95 98 97 96

P21 98 99 98 97 98 98 98 95 98 98 98 98 97 95 99 95

P22 97 97 95 98 97 95 97 97 98 96 97 98 98 97 97 96

P23 97 96 96 97 96 98 97 97 97 98 97 98 97 99 97 97

P24 96 95 98 94 97 98 94 98 96 93 97 99 98 98 94 97

P25 96 96 94 95 96 95 94 97 96 97 96 95 95 94 96 97

P26 94 94 95 95 97 98 94 98 94 95 93 92 95 97 95 94

P27 97 95 96 95 96 97 97 93 97 97 98 98 98 95 97 96

P28 96 93 95 92 96 95 94 94 97 96 98 98 96 96 98 97

P29 96 95 98 95 96 96 95 96 94 95 93 96 96 97 96 98

P30 97 95 97 96 97 97 96 96 96 93 94 97 96 94 95 96

P31 98 98 98 97 98 96 98 94 97 98 97 97 96 96 96 97

P32 97 97 97 98 98 96 97 96 97 98 96 97 97 97 96 96

P33 97 95 96 95 96 98 97 97 98 97 97 98 98 92 95 97

56

P34 95 95 96 93 93 98 95 97 97 96 98 95 96 98 94 97

P35 97 98 97 96 98 95 98 96 95 97 95 95 96 95 97 94

P36 96 95 96 97 96 98 96 98 97 96 96 97 97 93 96 97

P37 97 95 97 96 97 98 96 96 97 97 97 95 98 97 95 96

P38 96 96 95 96 96 93 97 98 96 97 95 98 97 93 97 98

P39 97 97 96 96 97 94 98 95 98 98 97 97 97 96 97 95

P40 96 92 95 96 95 96 97 97 96 96 96 95 96 98 94 98

P41 96 95 97 98 96 96 97 95 97 96 97 96 96 97 92 96

P42 96 93 95 96 95 96 98 96 96 95 96 99 97 96 96 97

P43 97 96 97 97 98 95 96 98 98 97 97 97 98 95 97 97

P44 97 98 96 96 97 96 97 96 97 97 97 96 97 94 98 96

P45 97 98 96 98 97 96 98 96 96 94 97 98 94 94 97 97

P46 98 97 98 97 98 96 97 98 98 97 97 96 97 98 97 97

P47 97 96 97 97 98 97 98 95 96 98 96 98 97 96 96 94

P48 98 97 98 97 98 98 98 97 97 96 97 98 97 97 98 98

P49 97 97 97 96 97 97 96 96 97 97 97 96 96 95 95 97

P50 98 97 96 97 97 96 95 96 97 95 98 97 98 98 96 94

P51 98 96 98 96 96 98 96 94 97 96 98 96 96 98 97 98

P52 96 98 96 96 93 96 98 100 97 98 97 92 97 93 99 98

P53 98 99 97 98 97 98 99 96 96 98 96 98 96 95 97 97

P54 97 94 96 98 95 96 97 97 97 97 96 96 96 96 98 98

Table 4.9: Test results in experiment 2, A ( 10/432 not recognized)

% E1 E2 E3 E4 E1 E2 E3 E4 E1 E2 E3 E4 E1 E2 E3 E4

P1 95 75 96 96 P28 94 88 93 93 P1 95 82 96 97 P28 96 92 99 97

P2 94 98 89 97 P29 91 69 96 96 P2 96 94 99 96 P29 93 66 94 90

P3 99 65 70 99 P30 89 78 94 97 P3 95 62 89 97 P30 91 43 91 96

P4 98 95 96 98 P31 98 88 97 98 P4 98 83 96 98 P31 96 95 87 95

P5 96 41 96 96 P32 98 74 89 97 P5 96 57 84 94 P32 98 93 97 96

P6 98 60 85 95 P33 93 93 98 96 P6 98 93 81 97 P33 97 96 97 99

57

P7 98 95 99 97 P34 95 80 99 95 P7 96 84 89 97 P34 94 76 89 95

P8 95 27 96 91 P35 98 90 96 97 P8 94 72 57 93 P35 97 24 96 96

P9 96 84 69 92 P36 90 75 93 96 P9 96 70 87 91 P36 93 77 94 96

P10 97 95 95 94 P37 96 87 94 97 P10 96 82 52 95 P37 97 95 97 97

P11 96 94 86 97 P38 98 59 89 94 P11 97 93 90 95 P38 94 39 97 95

P12 99 60 60 96 P39 95 82 96 97 P12 96 78 82 93 P39 97 67 93 97

P13 93 61 96 94 P40 89 77 97 96 P13 95 69 33 96 P40 93 28 98 97

P14 98 98 94 98 P41 97 88 96 96 P14 96 54 93 97 P41 93 74 96 97

P15 97 83 87 95 P42 96 79 98 97 P15 96 78 96 94 P42 96 70 95 97

P16 95 89 96 99 P43 94 81 96 98 P16 96 98 75 93 P43 96 95 94 97

P17 94 58 98 98 P44 98 77 87 96 P17 97 69 88 98 P44 97 36 78 97

P18 94 82 98 93 P45 98 87 95 97 P18 93 88 93 98 P45 96 81 81 96

P19 96 65 98 97 P46 97 82 94 98 P19 95 92 96 97 P46 97 80 94 98

P20 99 94 94 98 P47 98 77 96 97 P20 97 95 97 95 P47 95 85 84 96

P21 99 88 93 96 P48 98 90 96 98 P21 98 53 96 97 P48 97 83 91 97

P22 96 72 83 97 P49 97 59 93 98 P22 97 79 90 97 P49 93 91 76 98

P23 97 60 84 97 P50 92 95 79 98 P23 98 97 97 97 P50 89 80 93 93

P24 92 73 83 97 P51 98 87 86 97 P24 97 62 97 95 P51 96 83 81 94

P25 96 22 94 96 P52 98 62 76 95 P25 97 91 99 95 P52 99 92 97 97

P26 90 90 90 97 P53 99 91 96 98 P26 96 91 80 93 P53 97 82 96 95

P27 94 67 96 97 P54 93 48 83 96 P27 97 66 95 97 P54 98 90 90 96

Table 4.10: Test results of experiment2, B (59/432 Not recognized th=50%)

% E1 E2 E3 E4

E1 E2 E3 E4

E1 E2 E3 E4

E1 E2 E3 E4

P1 94 13 92 97 P28 92 51 85 95 P1 95 38 95 96 P28 94 77 78 96

P2 96 96 97 94 P29 65 73 71 95 P2 89 95 95 95 P29 95 36 89 95

P3 99 19 81 98 P30 51 78 93 97 P3 95 5 65 93 P30 59 27 92 97

P4 99 28 96 97 P31 96 89 84 98 P4 93 58 96 98 P31 90 68 90 97

P5 93 8 97 93 P32 99 50 76 96 P5 92 72 95 97 P32 93 47 95 96

P6 95 60 91 95 P33 95 12 97 97 P6 97 54 26 95 P33 96 90 97 98

58

P7 96 93 80 97 P34 95 23 86 96 P7 97 73 78 97 P34 97 97 96 99

P8 93 45 25 89 P35 93 39 96 97 P8 54 58 70 95 P35 95 31 94 96

P9 96 4 25 94 P36 73 28 95 95 P9 93 21 54 94 P36 84 15 84 96

P10 97 6 82 97 P37 93 74 95 97 P10 95 37 59 98 P37 90 42 95 97

P11 91 76 64 95 P38 96 39 58 97 P11 94 67 90 97 P38 79 5 89 94

P12 99 90 80 92 P39 86 4 95 97 P12 98 20 87 92 P39 93 26 95 97

P13 91 51 95 94 P40 96 28 96 98 P13 67 81 95 97 P40 75 37 89 96

P14 99 89 97 98 P41 81 66 91 93 P14 91 40 82 97 P41 91 81 79 96

P15 98 77 91 97 P42 92 46 96 95 P15 98 46 73 94 P42 97 6 98 97

P16 91 59 98 98 P43 96 72 75 97 P16 97 90 89 97 P43 95 49 83 98

P17 89 4 57 96 P44 97 33 84 97 P17 98 76 86 98 P44 96 4 53 97

P18 99 76 43 96 P45 90 66 87 97 P18 97 50 64 98 P45 94 24 68 96

P19 85 69 23 97 P46 96 44 75 98 P19 96 72 42 97 P46 98 69 75 97

P20 97 59 86 95 P47 97 59 50 97 P20 93 23 69 94 P47 95 37 86 97

P21 99 50 58 98 P48 94 75 83 97 P21 95 53 82 98 P48 86 69 85 98

P22 93 73 92 98 P49 90 45 61 97 P22 95 62 89 97 P49 74 43 83 97

P23 96 94 80 96 P50 90 93 94 97 P23 97 91 94 97 P50 96 92 96 97

P24 98 56 17 96 P51 97 85 93 98 P24 92 68 91 97 P51 97 67 68 96

P25 95 56 85 96 P52 96 25 74 95 P25 97 19 94 96 P52 98 33 76 96

P26 89 24 95 95 P53 99 32 94 98 P26 63 26 85 96 P53 98 9 88 95

P27 90 80 91 97 P54 89 28 83 96 P27 96 23 93 97 P54 98 83 29 95

Experiment 3 results

Table 4.11: Training results for experiment 3 (6/864 not recognized)

E1R E2R E3R E4R E5R E6R E7R E8R E1L E2L E3L E4L E5L E6L E7L E8L

P1 95 89 96 93 90 93 93 94 93 93 92 94 93 92 94 98

P2 91 90 91 93 93 93 89 88 93 92 94 95 93 97 89 94

P3 94 96 96 91 92 80 96 90 91 86 92 81 94 86 95 98

P4 95 94 94 95 94 91 92 97 93 94 93 92 93 85 95 94

59

P5 87 89 85 86 82 94 89 95 91 91 92 86 89 92 93 96

P6 87 91 88 85 86 80 73 84 91 92 93 80 91 84 91 88

P7 94 97 96 91 94 95 95 91 90 90 91 92 90 85 87 93

P8 92 93 91 90 89 92 92 81 90 91 78 88 83 91 88 91

P9 87 94 89 85 87 94 90 89 87 94 88 87 87 81 95 75

P10 92 94 88 93 90 90 93 89 92 95 95 88 90 81 93 92

P11 92 92 91 93 92 88 89 84 92 93 93 91 91 92 94 94

P12 93 90 87 87 92 87 94 96 87 93 85 90 85 83 95 92

P13 86 77 85 80 76 94 89 82 82 83 83 94 82 94 84 84

P14 93 93 94 94 94 92 93 92 92 88 94 93 92 90 94 96

P15 92 88 93 88 93 91 91 85 88 85 88 94 86 84 95 86

P16 90 90 95 87 90 90 93 92 95 79 93 86 96 95 95 86

P17 90 92 90 89 89 90 95 93 92 93 92 86 93 94 92 89

P18 91 87 92 95 93 92 91 94 91 88 95 92 91 90 92 97

P19 93 95 91 88 89 84 86 89 94 94 93 82 95 96 95 94

P20 92 96 93 89 93 93 96 94 90 91 91 92 88 96 94 92

P21 95 97 95 94 95 91 97 83 95 95 94 92 94 90 96 87

P22 92 92 91 95 93 91 95 87 94 94 92 92 93 89 87 96

P23 92 93 94 93 92 92 94 89 95 95 96 95 93 94 95 91

P24 89 87 93 89 94 93 83 89 90 82 91 97 92 93 82 96

P25 85 86 78 83 82 81 76 85 88 92 86 84 89 87 85 73

P26 90 79 89 88 89 94 88 91 86 81 84 89 86 91 79 80

P27 93 91 88 82 93 87 93 88 94 95 95 95 95 88 91 92

P28 87 85 83 80 86 88 84 87 92 88 93 91 93 91 93 92

P29 88 83 91 89 90 88 88 92 85 86 84 87 84 87 93 95

P30 90 89 92 91 90 87 89 93 89 83 87 92 88 70 87 91

P31 94 95 93 94 94 90 97 90 93 93 94 94 89 94 93 92

P32 93 94 93 94 94 91 93 96 94 95 92 90 95 90 94 90

P33 93 91 90 85 92 94 93 92 93 89 90 94 95 84 89 97

P34 85 85 88 83 80 89 89 91 91 90 94 90 90 95 87 86

P35 90 90 89 93 93 83 91 97 89 90 89 91 87 89 86 93

P36 93 85 92 89 89 91 88 87 92 88 91 90 92 82 87 92

P37 94 88 92 93 94 92 88 89 93 93 91 93 95 96 92 91

60

P38 89 89 86 92 88 89 90 92 92 94 91 95 95 73 94 90

P39 90 94 88 87 91 91 91 89 92 95 91 92 93 88 94 90

P40 90 87 88 89 90 91 88 95 87 90 87 92 88 88 91 96

P41 91 90 94 92 89 90 90 93 92 91 90 91 92 89 84 88

P42 88 91 92 86 89 84 94 88 91 93 89 92 89 95 94 93

P43 95 94 94 92 94 92 91 95 95 93 95 93 95 93 94 94

P44 93 91 90 92 93 92 92 89 92 94 90 93 92 92 94 92

P45 93 92 92 94 94 91 94 93 91 82 90 92 91 88 94 86

P46 95 96 95 91 96 91 97 95 94 92 93 89 92 95 91 92

P47 92 91 93 94 94 90 94 96 91 92 91 91 89 88 93 94

P48 96 94 95 94 96 96 96 94 94 91 95 94 93 95 91 95

P49 92 93 90 89 93 92 89 88 94 94 94 85 94 90 94 86

P50 93 87 86 93 92 93 88 90 94 87 95 92 91 96 85 93

P51 93 90 93 92 91 94 90 91 93 92 94 85 93 95 94 94

P52 90 92 91 93 86 88 95 97 90 91 88 86 88 82 94 95

P53 95 96 93 95 94 93 96 92 93 94 91 94 92 91 94 89

P54 91 86 89 91 89 94 91 89 94 91 93 92 92 91 94 96

Table 4.12: Test results, A (24/432 Not recognized, threshold = 0.5)

% E1 E2 E3 E4 E1 E2 E3 E4 E1 E2 E3 E4 E1 E2 E3 E4

P1 90 73 84 95 P28 82 62 91 81 P1 92 67 93 94 P28 86 96 91 92

P2 90 97 85 93 P29 83 57 78 89 P2 89 79 95 92 P29 78 56 88 79

P3 97 51 69 96 P30 87 75 78 89 P3 91 64 70 95 P30 82 69 88 89

P4 94 73 96 95 P31 95 94 92 94 P4 93 68 94 93 P31 90 48 78 86

P5 92 53 93 89 P32 95 90 87 93 P5 89 49 84 87 P32 95 83 83 93

P6 89 28 63 82 P33 83 92 89 90 P6 92 38 81 90 P33 92 95 95 95

P7 95 81 84 92 P34 79 86 84 84 P7 87 85 87 94 P34 89 91 82 93

P8 88 24 89 90 P35 94 58 90 88 P8 79 29 53 88 P35 91 40 81 89

P9 88 32 24 74 P36 82 67 89 91 P9 91 48 59 89 P36 83 76 87 93

P10 94 52 89 85 P37 92 85 93 94 P10 94 95 49 93 P37 89 76 93 93

61

P11 90 52 67 88 P38 92 40 63 86 P11 93 62 74 95 P38 93 66 95 92

P12 97 60 63 93 P39 93 85 82 91 P12 94 28 72 83 P39 93 69 86 91

P13 79 31 94 76 P40 83 68 95 91 P13 83 53 37 81 P40 79 36 92 90

P14 95 92 75 94 P41 90 33 85 93 P14 89 43 89 94 P41 93 81 90 92

P15 94 64 27 90 P42 89 77 92 90 P15 92 61 82 88 P42 88 57 95 92

P16 93 59 85 95 P43 91 84 95 97 P16 94 87 54 82 P43 89 90 81 94

P17 91 66 95 93 P44 93 58 82 92 P17 92 91 83 92 P44 92 21 84 94

P18 86 50 92 84 P45 95 61 87 94 P18 92 65 81 95 P45 92 53 42 87

P19 88 65 81 95 P46 92 89 97 96 P19 91 81 93 93 P46 91 59 84 92

P20 95 86 89 94 P47 94 58 81 91 P20 91 64 91 87 P47 93 76 83 85

P21 97 71 86 94 P48 94 87 94 96 P21 95 62 80 94 P48 92 90 83 95

P22 95 75 84 90 P49 87 36 77 90 P22 92 77 84 90 P49 83 77 79 94

P23 93 84 78 93 P50 88 74 32 93 P23 96 89 93 94 P50 79 66 62 87

P24 83 88 70 93 P51 95 78 90 93 P24 85 73 93 89 P51 90 79 79 92

P25 88 24 76 85 P52 91 73 64 85 P25 88 63 96 85 P52 95 90 83 90

P26 79 79 93 86 P53 97 77 84 97 P26 83 64 78 84 P53 94 64 80 91

P27 90 52 88 92 P54 87 64 90 88 P27 95 72 86 94 P54 94 97 85 94

Table 4.13: Test results B, (72/432 not recognized, threshold =0.5)

% E1 E2 E3 E4

E1 E2 E3 E4

E1 E2 E3 E4

E1 E2 E3 E4

P1 88 12 86 94 P28 74 35 51 87 P1 94 39 95 93 P28 82 82 79 90

P2 81 96 94 88 P29 87 77 84 84 P2 87 88 94 91 P29 90 46 85 79

P3 96 7 80 96 P30 61 46 92 91 P3 87 14 76 92 P30 53 13 77 89

P4 94 9 92 95 P31 93 68 85 92 P4 78 28 86 93 P31 80 54 78 91

P5 78 71 85 77 P32 91 61 69 92 P5 84 80 89 91 P32 80 33 93 93

P6 81 24 83 85 P33 92 40 94 92 P6 79 30 44 81 P33 82 89 94 93

P7 94 92 88 94 P34 91 70 61 86 P7 82 49 34 90 P34 60 75 84 93

P8 80 19 9 89 P35 90 2 92 89 P8 50 26 24 91 P35 78 33 93 86

P9 94 20 15 90 P36 64 64 71 89 P9 81 3 49 89 P36 78 27 90 92

P10 75 10 59 92 P37 89 63 90 94 P10 95 29 63 93 P37 88 15 90 94

62

P11 82 29 55 89 P38 85 35 31 89 P11 83 50 91 93 P38 96 53 85 95

P12 97 56 64 86 P39 85 3 73 90 P12 95 20 66 87 P39 94 17 93 92

P13 72 38 97 89 P40 95 4 93 92 P13 65 77 93 84 P40 58 70 64 89

P14 95 81 76 95 P41 80 14 97 90 P14 86 12 74 89 P41 87 68 61 91

P15 95 59 80 89 P42 93 9 81 86 P15 71 68 61 85 P42 90 6 96 91

P16 80 90 86 92 P43 90 80 85 96 P16 92 53 75 92 P43 88 64 86 96

P17 91 72 82 90 P44 86 26 81 91 P17 95 58 86 93 P44 88 24 52 92

P18 94 48 36 89 P45 81 21 77 92 P18 90 78 60 93 P45 88 5 80 92

P19 62 40 44 94 P46 90 63 67 96 P19 90 88 73 92 P46 90 61 56 91

P20 88 55 71 93 P47 92 88 92 92 P20 93 16 64 87 P47 96 15 64 92

P21 95 40 78 95 P48 93 46 86 94 P21 91 28 54 95 P48 84 57 88 94

P22 92 42 86 93 P49 74 42 39 94 P22 74 33 74 94 P49 63 31 79 92

P23 94 90 88 92 P50 45 37 57 91 P23 94 88 69 95 P50 82 65 47 93

P24 86 45 57 89 P51 95 86 88 91 P24 63 67 72 91 P51 90 60 71 91

P25 91 37 87 89 P52 92 78 65 85 P25 80 4 92 89 P52 95 62 72 84

P26 80 82 89 87 P53 98 78 89 96 P26 48 38 76 84 P53 91 34 65 91

P27 57 65 84 94 P54 96 74 57 85 P27 91 32 94 94 P54 96 35 66 94