chapter 2 - literature review 4 dec

44
CHAPTER 2 LITERATURE REVIEW 2.1 Introduction Recent studies on multibiometric systems are given in section 2.2. Section 2.3 explains the researches on single biometric systems using speech signal and lip image traits. Section 2.4 discusses the theory of support vector machine. Reviews on Fuzzy Inference System (FIS) are explained in Section 2.5. Finally, the summary is presented in section 2.6. 2.2 Literature Reviews on Multibiometric Systems Initially, researches on biometric systems have been focused on single modal verification which only consider single trait as biometric data. The systems only use a single modality to find the genuine person from the given database (Jain et al., 1999a). However, single biometric systems tend to obtain low performances when the data is corrupted due to noisy condition. Other than that, physical appearance and behavioral characteristics of a person tend to vary with time which also can affect 8

Upload: indahnyapelangi

Post on 23-Oct-2015

22 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Chapter 2 - Literature Review 4 Dec

CHAPTER 2

LITERATURE REVIEW

2.1 Introduction

Recent studies on multibiometric systems are given in section 2.2. Section 2.3

explains the researches on single biometric systems using speech signal and lip

image traits. Section 2.4 discusses the theory of support vector machine. Reviews on

Fuzzy Inference System (FIS) are explained in Section 2.5. Finally, the summary is

presented in section 2.6.

2.2 Literature Reviews on Multibiometric Systems

Initially, researches on biometric systems have been focused on single modal

verification which only consider single trait as biometric data. The systems only use

a single modality to find the genuine person from the given database (Jain et al.,

1999a). However, single biometric systems tend to obtain low performances when

the data is corrupted due to noisy condition. Other than that, physical appearance and

behavioral characteristics of a person tend to vary with time which also can affect

biometric system performances (Kittler et al., 1997b). One of the solutions to

overcome this problem is by implementing multibiometric systems. Multibiometric

systems combine multiple traits (i.e. speech, iris and fingerprint) in order to improve

the systems recognition accuracy when one of the traits is corrupted (Rowe et al,

2007). Multibiometric refers as an extension of a single biometrics in which

information from multiple sources such as sensors, units, samples, algorithms and

traits are combined. Further reports on multibiometric systems have been reviewed in

(Ross et al., 2004) and (Ross et al., 2007).

8

Page 2: Chapter 2 - Literature Review 4 Dec

2.2.1 Advantages and Disadvantages of Multibiometric Systems

According to Rowe et al. (2007), multibiometric systems are capable to solve

the non-universality problem faced by single biometric systems. For example, if a

mute person who is unable to provide information required by the speaker

verification systems, the person can aid the problem by using other biometric traits

such as fingerprint, iris or face. Besides, multibiometric systems can avoid problems

caused by noisy data where the information obtained is not sufficient for decision-

making. The system can implement the data from other traits which provide

sufficient information in order to enable the decision-making. Another advantage of

multibiometric systems is where the spoof attacks can be avoided since the systems

require more than one trait which is harder for the imposter to mimic the enrollment

speakers.

However, one of the disadvantages faced by multibiometric systems is that it

requires more sensors which contribute to higher implementation cost compared to

single biometric systems. In addition, multibiometric systems require user to interact

with more than one sensor. For example, multibiometric system using microphone

and fingerprint scanner may increase the user inconvenience since a user needs to

provide information for microphone as well as to touch the fingerprint scanner.

Hence, more computation, memory and storage are required for this purpose.

Furthermore, the operating times during enrolment and verification process are also

increased (Rowe et al., 2007).

2.2.2 Taxanomy of Multibiometric Systems

Multibiometric systems can be classified into five systems i.e. multimodal,

multi-sample, multi-algorithm, multi-sensor and multi-instance systems as shown in

Figure 2.1.

9

Page 3: Chapter 2 - Literature Review 4 Dec

Figure 2.1: Scenarios in multibiometric systems

a. Multimodal systems: multimodal systems extract multiple biometric

information from many modalities such as speech, lip and face for

verification of individuals. The multimodal systems are highly reliable

especially if one of the traits has insufficient information. However, cost of

developing these systems are higher due to more sensors needed to extract the

traits’ information. Study on multimodal systems using face and speech

modalities has been reported by Brunelli and Falagivna (1995). According to

Kittler et al., (1998), the multimodal systems performances have been

improved by combining three biometrics i.e. frontal face, face profile and

voice using sum rule combination scheme. In another research, a multimodal

system implementing three different traits i.e. fingerprint, face and finger vein

has been discussed in Hong et al. (1999).

b. Multi-sample systems: Multi-sample systems use multiple samples extracted

from a single trait which obtained from a single sensor. The scores are

extracted from each sample by applying the same algorithm in order to obtain

an overall recognition results. The benefit of using multiple samples is to

10

Page 4: Chapter 2 - Literature Review 4 Dec

avoid poor performance due to the bad properties of sample if only one

sample is used during the process. Research by Samad et al. (2007) proposed

multi-sample fusion schemes in order to increase the system performances.

This fusion scheme computes the score from each sample using maximum,

minimum, median, average and majority vote operator. According to Suutala

and Roning (2005), the combinations of multi-samples have improved the

performances of footstep profile-based person identification. This study

employed multiple footsteps from each person and the scores were fused

using simple product and sum rules fusion schemes.

c. Multi-algorithm systems: multi-algorithm systems are a combination of

output obtained from multiple methods such as classification algorithms

or/and feature extraction for the same biometric data (Ross and Jain, 2007).

The outputs are combined to obtain an overall recognition result. Hence,

increase the systems performances. The advantage of multi-algorithm is that

no multiple sensors are required which is very cost effective. However, the

system computation will be complicated due to many feature extractions and

matching modules are verified during the process. Studies on multi-filter bank

approach for speaker verification based on genetic algorithm have been

discussed by Charbuillet et al. (2007b). This study proposes a feature

extraction system based on the combination of three feature extractors (i.e.

MFCC, LPCC and LFCC) adapted to the speaker verification task. Results

proved that the proposed method improves the system performances. In

addition, subspace algorithms such as PCA, Fisher Linear Discriminant

(FLD) and ICA have been applied for palm print and face separately in order

11

Page 5: Chapter 2 - Literature Review 4 Dec

to determine the best algorithm performance. This study can be found in

Imran et al. (2010).

d. Multi-sensor systems: Multi-sensor systems implement multiple sensors to

capture single biometric trait of an individual. Marcialis and Roli (2004)

reported that the multi-sensor systems can perform better than traditional

fingerprint matchers using a single sensor. According to Lee et al. (2004),

images of a subject are captured using multiple 2D cameras. Next, extraction

of a person’s face using an infrared sensor and visible-light sensor has been

illustrated in Kong et al. (2005). Subsequently, multi spectral cameras were

used to extract the images of iris, face and finger have been explained in

Rowe and Nixon (2006).

e. Multi-instance systems: For multi-instance systems, the biometric

information is extracted from the multiple instances of a single biometric

trait. As an example, Prabhakar and Jain (2000) proposed the used of the left

and right index finger and iris of an individual. A study on multi-instance

speech signal data fusion by evaluating the multi-instance of speech signal

(i.e. zero, seven and eight) has been discussed in Ramli et al. (2010).

According to Ramli et al. (2011), combination of three speech modality

subsystems from different verbal zero, seven and eight multi-instance were

proposed to overcome the limitations faced by single modal system.

2.2.3 Levels of Fusion of Multibiometric Systems

The levels of fusion of multibiometric systems can be classified into fusion

before matching and fusion after matching as illustrated in Figure 2.2. Fusion before

matching-pattern process is known as pre-classification while fusion after pattern

matching process is known as post-classification. For fusion before pattern matching,

12

Page 6: Chapter 2 - Literature Review 4 Dec

the fusion process is computed at sensor and feature levels while fusion after

matching-pattern, the fusion process is done at match score, rank and decision levels

(Ross and Jain 2007).

Figure 2.2: Level of fusions

a. Fusion before matching

i. Sensor level fusion: Sensor level fusion is a combination of raw data

extracted from the sensor as displayed in Figure 2.3. According to

Iyengar et al. (1995), the system performances may be affected due to

contaminate raw data which results from many factors such as

background clutter and non-uniform illumination. As discussed by

Singh et al. (2005), two types of conditions i.e. data from a single trait

obtained from multiple sensors and data from multiple snapshots of a

single biometric trait extracted from a single sensor can be performed

during sensor level fusion.

13

Page 7: Chapter 2 - Literature Review 4 Dec

Figure 2.3: Sensor level fusion process flows

ii. Feature level fusion: Feature level fusion is a combination of different

feature vectors extracted from multiple biometric sources into a single

feature vector as illustrated in Figure 2.4. Feature normalization and

feature selection have been executed during the process. According to

(Jain et al., 2005), the feature normalization is implemented to modify

the location and scale of feature values via a transformation function

where can be done by using appropriate normalization schemes. As an

example, the min-max technique and median scheming have been

used for hand and face biometric traits. Next, feature selection is

executed in order to reduce the dimensionality of a feature vector.

This process has the ability to improve the matching performance. As

stated in (Kumar and Zhang, 2005), feature selection algorithms such

as Sequential Backward Selection (SBS), Sequential Forward

Selection (SFS) and Partition About Medoids have been studied.

However, the feature level fusion is hard to perform due to the joint

feature set extracted from different biometric sources may not be

linear and incompatible (Ross and Jain, 2004). Therefore, more

researchers focused on other types of fusion schemes such as score

level fusion and decision level fusion.

14

Page 8: Chapter 2 - Literature Review 4 Dec

Figure 2.4: Feature level fusion process flows

b. Fusion after matching

i. Rank level fusion: Rank level fusion is a combination of identification

ranks obtained from multiple unimodal biometrics as shown in Figure

2.5. A novel approach to improve biometric recognition using rank

level fusion has been reported in Bhatnagar et al (2007). This paper

used the rank level fusion which improved the system performances.

Rank level fusion using fingerprint and iris biometric has been

discussed in Radha and Kavitha (2011). The experimental results have

revealed better performances of the proposed rank level fusion in

multimodal biometrics system.

Figure 2.5: Rank level fusion process flows

ii. Score level fusion: Score level fusion is a combination of match

outputs from multiple biometrics in order to improve the matching

performances (Jain and Ross, 2004). This approach is illustrated as in

Figure 2.6. The fusion of this level is widely applied in multibiometric

systems due to its simplicity. Moreover, the matching scores consist

of sufficient information which enables the systems to distinguish the

authentic users from the imposter users (He el al., 2010). However,

the combination process may be defected due to degrade biometric

performance. Score level fusion can be grouped into three schemes

15

Page 9: Chapter 2 - Literature Review 4 Dec

i.e. density-based schemes, transformation-based scheme, and

classifier-based scheme in order to overcome this limitation (Ross and

Jain, 2007). In density-based scheme, a training set of the genuine and

imposter match scores is estimated using a joint density function.

Next, the posterior probability of observing genuine (or imposter)

class are defined from the Bayes formula. However, the density-based

fusion scheme requires large training samples in order to develop

accurate system performances. Therefore, it is not suitable in most of

the multimodal systems because of the time consuming and high cost

factors. The transformation-based scheme is applied during the score

normalization process which requires simple normalization technique,

i.e. min-max normalization, z-score and tanh-estimator in order to

transform the score into the same domain. For transformation-based

fusion, the match score is directly combined using simple fusion

operators such as sum rule, product rule, min rule and max rule

techniques (Parviz and Moin, 2011). For the classifier-based scheme,

the matched scores extracted from each biometric source are used as

inputs to a trained pattern classifier such as SVM, HMM and ANN.

Next, the input score is classified by the classifier in order to identify

the genuine and imposter class (Nandakumar et al., 2007).

16

Page 10: Chapter 2 - Literature Review 4 Dec

Figure 2.6: Match score level fusion process flows

iii. Decision level fusion: Each classifier in decision level fusion operates

under a binary hypothesis which applies a threshold on the match

score and renders its decision regarding the presence (=1) or absence

(=0) of a genuine individual. The decisions from multiple classifiers

are fused in order to obtain the final decision as shown in Figure 2.7.

Few methods such as “AND” and “OR” rules, majority voting,

weighted majority voting and Bayesian decision fusion have been

implemented in order to combined the distinct decisions to a final

decision (Ross and Jain, 2007). Messer and Kittler (2000) discussed

the data and decision level fusion of temporal information for

automatic target recognition (ATR). An adaptive ATR system which

decides how to best distinguish the target from a particular

background has been proposed in this study. In decision level fusion,

Kahler and Blasch (2011) discussed the used of decision levels fusion

in High Range Resolution (HRR) radar and Synthetic Aperture Radar

(SAR). Results proved that the decision level fusion is able to enhance

identification performance.

17

Page 11: Chapter 2 - Literature Review 4 Dec

Biometric Information + Quality Information Class Level

Genuine or imposter user

Subjective Quality (Q)Objective Quality (q)Additional Knowledge i.e. Device (d)

Biometric DataFeature representationScore (y)

Figure 2.7: Decision level fusion process flows

2.2.4 Quality Based Fusion Systems

Quality measures are measurements to identify the degree of excellence of

biometric samples to some predefined criteria which may influence the system

performance. This study implements the quality measures into the biometric systems.

Figure 2.8 described the general theory of incorporating quality measures in

biometric systems.

Figure 2.8: General theory of incorporating quality measures

The quality information can be grouped into two classes, i.e. subjective

quality (Q) and objective quality (q), respectively. The subjective quality (Q) is

derived from human judgement, manual (not computable) and observable only

during training while the objective quality (q) is derived from biometric sample,

18

Page 12: Chapter 2 - Literature Review 4 Dec

automatic and computable as well as observable during both training and testing.

Examples for subjective and objective quality for face and iris are shown in Figure

2.9 and 2.10, respectively.

Good quality or bad quality?

Uniform background High spatial resolution High colour resolution

Cluttered background Low spatial resolution Low colour resolution

q So, background uniformity, spatial resolution and colour resolution are potential quality measures for face

Figure 2.9: Example 1 (face) for subjective and objective quality (Poh et al., 2010)

Good quality or bad quality?

i.Good quality, ii. Small iris area, iii. Displaced contact lens and iv. Blurred

q So, iris area, contact lens and image focus are potential quality measures for iris

Figure 2.10: Example 2 (iris) for subjective and objective quality (Poh et al., 2010)

Research on discriminative multimodal biometric authentication based on

quality measures has been illustrated in Fierrez et al. (2005). Chen et al. (2005) has

implemented the fingerprint quality indices for predicting authentication

performance in their studies. Subsequently, incorporating image quality multi-

algorithm fingerprint verification has been discussed in Fierrez et al. (2006).

19

Q

Q

Page 13: Chapter 2 - Literature Review 4 Dec

2.3 Literature Reviews on Single Biometric Systems

2.3.1 Speech Signal Trait

The technology of Automatic Speech Recognition (ASR) has progressed

greatly over the past few years.  During 1939, studies on automatic speech

recognition and transcription have started by the ATT&T's Bell Labs. In 1960,

Gunnar Fant created a model which can describe the physiological components of

speech signal using the x-rays analysis. Each speech signal produces different phonic

sounds which can be used to identify the speakers. In 1970, the Fand model has been

expanded by Dr. Joseph Perkell by adding the tongue and jaw to the model (John et

al, 2003). As mentioned by Haberman and Fejfar (1976), the National Institute of

Standard and Technology (NIST) developed the NIST Speech Group in the mid

1980s to study the uses of speech processing techniques. Studies on ASR have been

further discussed in Campbell (1997) and Reynolds (2002). Features and techniques

for speaker recognition can be found in Singh (2003).

Research on speaker recognition using Mel Frequency Cepstrum Coefficient

(MFCC) features has been implemented in Hazen et al. (2004) and Hazen (2006).

Linear Prediction Coding (LPC) features has been discussed in Atal (1974) and Furui

(1981a). Furui (1981b) has used the fixed Japanese utterences features for the

speaker recognition system. Other features that have been suggested are Perceptual

Linear Predictive (PLP) coefficient which has been explained in Xu (1989) and Line

Spectral Pair (LSP) frequencies which has been studied in Liu (1990). The LPC

feature is modified to PLP coefficient which based on human perception and

physiological effect sound while LSP coefficient which based on formant

bandwidths and locations. Partial Correlation (PARCOR) is also modified from the

LPC feature which has been discussed in Atilli (1988). Researches by Dupont and

20

Page 14: Chapter 2 - Literature Review 4 Dec

Luettin (2000) and Heckmann et al. (2002b) have implemented noise-robust

RASTA-PLP features for the speaker recognition system. Linear Discriminant

Analysis (LDA) feature is a combination of LPC coefficient and dynamic LPC

coefficient in order to reduce the dimension of speech vector space has been reported

in Hai and Joo (2003). The UBM_GMM cepstral features, prosodic statistics and

pronounciation modelling has been explained in Campbell (2003). Study by Jiang et

al. (2003) has proposed the wavelet packet strategy during the feature extraction. In

another research, filter bank features has been discussed in Ravindran et al. (2003).

For the pattern matching process, few techniques such as Euclidean,

Manhattan Distance, Dynamic Time Warping (DTW), Vector Quantization (VQ),

Hidden Markov Model (HMM), Gaussion Mixture Model (GMM), Artificial Neural

Network (ANN) and Support Vector Machine (SVM) have been proposed. For text

independent recognition, speaker specific VQ codebooks or the more

advanced GMM have been used regularly. Furthermore, Euclidean and Manhattan

Distance are simpler techniques compared to more advanced technique such as

DTW, VQ and HMM. DTW is an algorithm for measuring similarity between two

sequences which may vary in time or speed. DTW approach may be a better choice

for a real-world speaker recognition system if the amount of available training data is

not sufficient. DTW has been used in speech signal processing (Rabiner & Juang

1993), manufacturing (Gollmer & Posten 1995), gesture recognition (Gavrila &

Davis 1995), medicine (Caiani et. al 1998) and robotics (Schmill et. al 1999).

In another research, the vector quantization (VQ) is another technique for

pattern matching process. The training data has been used to form a speaker’s

codebook. During the recognition stage, the test data is compared to the codebook of

each reference speaker and a measure of the difference has been used to make the

21

Page 15: Chapter 2 - Literature Review 4 Dec

recognition decision. This model used a vector quantized codebook, which is

generated for a speaker by using the training data. The VQ technique has been

discussed in Linde et al. (1980) and Soong et al. (1985), respectively.

A model that is widely used for modeling of sequences is the Hidden Markov

Model (HMM). It provides more flexibility and produce better matching score. In

speech recognition, HMMs have been used for modeling observed patterns from

1970s. However, the system performances are quite slow compared to other

methods. Many researchers i.e. Fu (1980) and Russell and Moore (1985) have

published a large number of papers, which present HMM as tool for use on these

practical problems. The HMM algorithm and its implementation has been described

by Rabiner and Juang (1986).

Next, GMM has also been used for pattern matching. This method has been

the most successful because of many factors such as the high-accuracy recognition,

the probabilistic framework, and the remarkable capability to model irregular data.

This characteristic makes it very suitable to have a smooth estimation of speaker’s

acoustic variability Studies on GMM based speaker recognition on readily available

databases has been discussed in Wildermoth and Paliwal (2003). Audio signal

processing using GMM has been implemented by Kumar et al. (2010).

Speech recognition by self-organizing feature finder using ANN has been

stated in Lerner and Deller (1991). ANN based on multi-layer perceptions or radial

basic function has been trained to discriminate between enrolled speaker and non-

enrolled speakers. Consequently, phonetic speaker recognition using SVM has been

explained in Campbell (2003). Study by Solomonof et al (2005) has used SVM in

order to derive fully non-linear channel compensations. Another method i.e.

correlation filters has been explained in Kumar (1992) and Alfalou et al. (2010). Face

22

Page 16: Chapter 2 - Literature Review 4 Dec

verification using correlation filters can be found in Savvides et al. (2002) while

lower face verification centered and lip using correlation filters have been discussed

in Samad et al. (2007) and Ramli et al. (2007). Synthetic Discrimination Function

(SDF) and Equal Correlation Peak SDF (ECP SDF) based techniques have been

proposed to overcome the problem when matched filters tend to drops due to changes

of scale, rotation and pose of reference images. Research on SDF has been

implemented in Rehman et al. (2005) while ECP SDF has been explained in

Alkanhal (2006).

Advanced correlation filter namely MACE and UMACE have also applied in

image processing which mostly used in authentication and identification process. The

advantages of MACE and UMACE are that they are easy to be implemented and do

not require large number of training images as reported in Tahir et al. (2005), Ramli

et al. (2008) and Ghafar et al. (2008).

Pattern matching techniques using probability density function has been

discussed in Schwartz et al. (1982). The K nearest neighbours method combines the

strengths of the DTW and VQ methods where it keeps all the data obtained from

training phase. Higgins et al. (1993) has proposed the used of K nearest neighbours

technique during the pattern matching process. Another technique i.e. verbal

information verification (VIV) has also been discussed by Li et al. (2000).

2.3.2 Lip Reading Trait

According to Petajan and Brooke (1988), humans use lip reading to enhance

speech recognition especially when the signal is degraded by noise or hearing

impairment. A number of techniques have been reported to extract mouth features for

visual lip reading. Report by Brooke (1996) has suggested that visible facial gestures

could be exploited to enhance speech intelligibility in automatic systems. This paper

23

Page 17: Chapter 2 - Literature Review 4 Dec

used image coding and compression for feature extraction. Active contour models

have been used in Kass et al. (1987) in order to detect lip contours while Chiou and

Hwang (1994) and Li et al. (1995) have used the eigensequence approach which is

often used in facial expression recognition.

There are few types of lip features for automatic lipreading which can be

divided into two groups i.e. high-level shape (lip contour) and low-level appearance

(image pixel). Common techniques that used for lip contour are B-Spline, eigenlip,

snakes and Active Shape Model (ASM). Eigenlip features for visual representation

have been reported in Bregler and Konig (1994). Statistical chromaticity for lip

tracking with B-spline has been explained in Sanchez et al. (1997). Chan et al. (1998)

has used the cubic B-spline to model the shape contour. Sum et al. (2001) has applied

ASM to extract the point-based lip contour. The lip contour evolution becomes better

controlled and fast convergence is achieved. The advantages of geometry features

have been evaluated by Broun et al. (2002). This study has used polynomial based

classifier during the experiments. Lip tracking using pattern matching snakes has

been reported in Barnard et al. (2002). Studies on geometry features such as height,

width and angel of speaker have been reported in Kaynak et al. (2004). A combined

balloon pressure and image gradient snake with adaptive colour thresholding has

been described in Kuo et al. (2005). Result has showed a high success rate for lip

fitting. Jang (2007) has used snakes and ASM to extract the lip contour.

The second approach which is low-level appearance (image pixel) has been

further discussed in Bradley (2002). ROI is a certain parts of an image which contain

the most informative features. The ROI image can be transformed into another

feature using Discrete Cosine Transform (DCT), Principle Component Analysis

(PCA) and Linear Discriminant Analysis (LDA). Face recognition system based on

24

Page 18: Chapter 2 - Literature Review 4 Dec

PCA and LDA has been presented in Marcialis and Roli (2002). Nhat and Lee (2005)

have proposed new PCA-based method which improves the performance of the

traditional PCA and two-dimensional PCA (2DPCA) approaches. Introduction of a

new algorithm for shape from focus (SFF) based on DCT and PCA have been

reported in Mahmood et al. (2008). Akrouf et al. (2009) has presented a hybrid

method combining principal components analysis (PCA) and the discrete cosine

transform (DCT).

Between these two approaches, the appearance based has been proved to

outperform the lip contour based. Moreover, the lip contour based requires expensive

computation. Few methods such as HMM, GMM and polynomial classifier is used

for pattern matching process.

2.4 Support Vector Machine

SVM is a supervised learning method that analyzes data and recognizes

patterns, widely used for classification and regression analysis. SVM is a binary

classifier, which models the decision boundary between the two classes as a

separating hyper plane. SVMs have also been applied to a wide variety of biological

applications as well as the automatic classification of microarray gene expression

profiles (Campbell, 2003).

The foundations of Support Vector Machines (SVM) have been developed as

been discussed in (Vapnik, 1995) and are gaining popularity due to many attractive

features, and promising empirical performance. Theory regarding SVM has been

further explained in (Gunn, 1998).

According to Gunn (2005), the solution of linearly separable case is given as

follows. Consider a problem of separating the set of training vectors belong to two

separate classes,

25

Page 19: Chapter 2 - Literature Review 4 Dec

D= {( x1 , y1 ) ,… .. ,(xL , y L)}, x∈Rn , y∈ {−1,1 }(2.1)

with a hyperlane,

⟨ w , x ⟩+b=0(2.2)

where w and b characterizes the direction and position in space, respectively, and w

is normal to the plane. For each direction, w has the same distance from the nearest

points from each class and the margin is twice the distance. Decision boundary in

support vector machine is presented in Figure 2.11. The hyperplane is at its optimal

when set of vectors from both classes are able to separate without any error and the

distance between the closest vectors to the hyperplane is maximal. The optimal

hyperplane, w0 , is a linear combination of a small subset of data, xs, s∈ {1 ,……, N }.

Figure 2.11: Decision boundaries in support vector machine

According to Vapnik (1995), the canonical hyperplane is derived by

mini

|⟨w , x i ⟩+b|=1(2.3)

where the parameters are w and b.

In canonical form, the hyperplane must satisfy the following constraints,

y i ⌊ ⟨w , x i ⟩+b ⌋≥ 1 , i=1 ,…… L(2.4 )

where the distance d(w, b; x) of a point x from the hyperplane (w, b) is

26

Page 20: Chapter 2 - Literature Review 4 Dec

d (w , b , x )=|⟨ w , x i ⟩+b|

‖w‖(2.5)

By maximizing the margin, ρ, subject to the constraints (2.4) where the margin is

given by

ρ (w , b )= minxi : y i=−1

d (w , b ; x )+¿ minx i : yi=−1

d (w , b ; x )= 2‖w‖

(2.6)¿

The hyperplane that optimally separates the data is written as

∅ (w )=12‖w‖²(2.7)

which is equivalent to minimizing an upper bound on VC dimension. VC dimension

is a scalar value that measures the capacity of the learning function. The solution to

the optimization problem (2.7) with constraints of (2.4) is given by the saddle point

of the Langrange functional (Langrangian)

∅ (w , b ,α )=12

‖w‖² - ∑i=1

L

αi( y i [ ⟨ w , x i ⟩+b ]−1) (2.8)

where a iare the Langrange multipliers. The Langrangian has to be minimized with

respect to w, b and minimized the respect to α ≥ 0. Equation (2.4) is the transformed

to its dual problem. The solution of the linearly separable case is given by,

α ¿=arg minα

12∑i=1

L

∑i=1

L

αi α j y i y j ⟨ x i . x j ⟩−∑k=1

L

αk (2.9)

with constrains,

0 ≤ αi ≤ C , i=1 ,……, L∧∑j=1

L

α j y j=0(2.10)

The Langrange multipliers α iare then determined by solving (2.9) with constraints

(2.10) and the optimal separating hyperplane is defined as

w ¿=∑i=1

L

αi y i x i(2.11)

27

Page 21: Chapter 2 - Literature Review 4 Dec

b¿=−1

2(w¿ , xr+xs )(2.12)

xr and xs are any support vector from each class which satisfying,

α r , α s>0 , y yr=−1 ys=1(2.13)

Support vector (SV) is defined as x i with non-zero α i. As stated in (2.11), w is a

linear combination of a small number of data points that inferring the hyperplane is

determined by a small subset of the training set. The hard classifier for linearly

separable case is gives as

f ( x )=sgn ( ⟨w ¿ , x ⟩+b ) (2.14)

while a soft classifier for linearly interpolates is given as

f ( x )=h ( ⟨ w¿ , x ⟩+b )(2.15)

where

h ( z )={−1 : z<−1z :−1 ≤ z ≤1+1 : z>¿1

(2.16)

For the case where the training data is linearly non-separable, non-negative variables,

ξ i≥ 0and a penalty function,

Fσ (ξ )=∑i

ξiσ σ>0 (2.17)

where the ξ i are a measurement of the misclassification errors. In order to minimize

the classification error as well as minimizing the bound on the VC dimension of the

classifier, the constraint in equation (2.4) is modified to equation (2.18).

y i ⌊ ⟨w , x i ⟩+b ⌋≥ 1−¿ ξ i ,i=1 , …. L¿)

The generalized optimal separating hyperplane is determined by the vector w by

minimizing the functional,

ɸ (w , ξ )=12‖w‖2+C∑

i

ξi(2.19)

28

Page 22: Chapter 2 - Literature Review 4 Dec

subject to the constraints (2.18). The saddle point of Lagrangian defined from

optimization problem of (2.19) under constraints (2.18) is defined as

ɸ (w ,b ,α , ξ , β )=12‖w‖2

+C∑i

ξ i−∑i=1

L

α i ( y i [wT x i+b ]−1+ξ i)−∑j=1

L

βi ξ i

(2.20)

where α i and β iare the Langrange multipliers which has to be minimized with respect

to w, b, x, and maximized with respect to αand β. For linearly separable case,

equation (2.20) is transformed to its dual problem. The linearly non-separable case is

given by,

α ¿=arg minα

12∑i=1

L

∑i=1

L

αi α j y i y j ⟨ x i . x j ⟩ K ⟨ x i . x j ⟩−∑k=1

L

α k (2.21)

with constraints,

0 ≤ αi ≤ C , i=1 ,……, L∧∑j=1

L

α j y j=0(2.22)

Figure 2.12 shows the transformation of input vector from input space to

feature space which overcomes the drawback occurs from the inappropriate linear

boundary. According to Gunn (2005), the SVM creates a separating hyperplane in

the higher dimensional space by choosing a non-linear mapping a priori. During the

process, the mapping is achieved by replacing the value of dot products between the

data points in input space with the value that results when the same dot product is

carried out in the feature space. The dot product in the feature space is described by

the kernel as some function of two data points in input space.

29

Page 23: Chapter 2 - Literature Review 4 Dec

Figure 2.12: Transformation from input space to feature space

Commonly, the kernel functions are certain sigmoid function, polynomials and radial

basic functions. The polynomial kernel is implemented in this study. Hence, the

optimization problem of (2.21) becomes

α ¿=arg minα

12∑i=1

L

∑i=1

L

αi α j y i y j ⟨ x i . x j ⟩ K ⟨ x i . x j ⟩−∑k=1

L

α k (2.23)

with constraints,

0 ≤ αi ≤ C , i=1 ,……, L∧∑j=1

L

α j y j=0(2.24)

The kernel function that performs the non-linear mapping into feature space is

described as K ⟨ x i . x j ⟩ in the equation (2.23). The polynomial kernel is defined as

K ⟨ x i . x j ⟩ = ∅ (x i)T ∅ (x j) = (γ xi

T x j – r)d (2.25)

where γ > 0 and γ , rand d are kernel parameters. Then, the Langrange multipliers are

executed by solving equation (2.23) with the constraints (2.24). The hard classifier

implementing the optimal separating hyperplane in the feature space is defined by

f ( x )=sgn¿

where

⟨ w¿ , x ⟩=¿ ∑i=1

L

αi y i K ( x i , x )(2.27)

30

Page 24: Chapter 2 - Literature Review 4 Dec

b¿=−12∑i=1

L

αi y i [ K (x i , xr)+K (x i , xr)](2.28)

Equation (2.28) can be executed using all the SV on the margin in order to gain

stability.

2.5 Fuzzy Logic

Fuzzy logic can accommodate the differences of human languages and logics

which provide an intuitive method for executing systems in human terms. It

automates the conversion of those system specifications into effective models. The

fuzzy logic has an ability to add human-like subjective reasoning capabilities into

machine intelligences as described in Prade and Dubois (1999). According to Vasuhi

et al. (2010), the fuzzy logic decision-making is approximately the same with the

human decision-making.

Many researches on fuzzy logic have been reported. For example, enhanced

password authentication through fuzzy logic has been discussed in De Ru and Eloff

(1997). Report on multimodal biometric system using fuzzy logic decision fusion has

been discussed in Lau et al. (2004). Next, a generic multimodal framework for

handling error patterns using fuzzy logic has been studied in McGlaun et al. (2004).

In another research, works on a fuzzy multimodal technique in order to guarantee the

desired level of security has been reported in Azzini et al. (2007). Moreover, study

on fuzzy fusion in multimodal biometric system has been reported in Conti et al.

(2007). Hui et al. (2007) has implemented the fuzzy logic decision fusion for the

adaptive weight estimation in multibiometric verification. Furthermore, fuzzy logic

has been used in order to develop an efficient multimodal biometric person

authentication system (Vasuhi et al. 2010).

31

Page 25: Chapter 2 - Literature Review 4 Dec

Fuzzy inference systems (FIS) are one of the most commonly used

applications of fuzzy logic and fuzzy sets theory (Zadeh, 1965). Basically four steps

are required for FIS i.e. fuzzification of input variable, rule evaluation, aggregation

of the rule outputs and defuzzification.

Step 1: Fuzzification

According to Negnevitsky, (2005), fuzzification is a process of translating the crisp

inputs of variable into linguistic variable using the membership functions. The input

is a crisp value and the output is a fuzzy degree of membership in the qualifying

linguistic set. The crisp inputs are identified to which it belongs to via the

membership functions.

Step 2: Rule Evaluation

The fuzzy output is computed by applying reasoning using IF-THEN rule based.

During the rule evaluation, the fuzzified inputs are applied to the antecedents of

fuzzy rules. The output which is a single number is obtained by applying the AND

and OR operators to the multiple antecedents. In fuzzy logic toolbox, there are two

built-in AND methods which are min (minimum) and prod (product) while two built-

in OR methods which are max (maximum) and probor (probabilistic OR).

Step 3: Aggregation

Aggregation is the process of unification of the outputs of all rules. The membership

functions for all rules are scaled and combined into a single fuzzy set. The

aggregation’s inputs are the list of scaled membership functions and the output is one

fuzzy set for each output variable.

Step 4: Defuzzification

32

Page 26: Chapter 2 - Literature Review 4 Dec

Defuzzification is the last step for the FIS process. The input for the defuzzification

process is the aggregate output fuzzy set and the output is a single number which is a

crisp number. This study applied the centroid technique for the defuzzification

process. This method determines the center of mass of the results and calculated the

crisp value.

2.5.1 Mamdani-type FIS versus Sugeno-type FIS

Two types of fuzzy inference system (FIS) i.e. Mamdani-type FIS and

Sugeno-type FIS are implemented in this study. Mamdani-type FIS is widely used

for capturing expert knowledge which is able to describe the expertise in more

human-like manner and more intuitive. However, Mamdani-type FIS may face a

substantial computational burden. On the other hand, Sugeno-type FIS is

computationally efficient and works well with adaptive techniques and optimization,

which makes it very useful especially in non linear systems. The main difference

between Mamdani-type FIS and Sugeno-type FIS is the way the crisp output is

generated from the fuzzy inputs. Mamdani-type FIS uses the defuzzification

technique to compute the fuzzy output while Sugeno-type FIS uses weighted average

technique to compute the crisp output. Sugeno-type FIS is less time consuming

compared to Mamdani-type FIS since the weighted average technique replace the

time consuming defuzzification process. In another research, Mamdani-type FIS has

output membership functions whereas Sugeno-type FIS has no output membership

functions (Arshdeep and Amrit, 2012).

2.6 Summary

In short, this chapter discussed the studies on multibiometric systems. This

chapter also reviewed the studies on single biometric system using speech signal and

lipreading traits. Subsequently, the quality based multibiometric systems are

33

Page 27: Chapter 2 - Literature Review 4 Dec

reported in this chapter. Moreover, this chapter discussed the support vector machine

implemented in this study. Finally, fuzzy inference systems (FIS) are

explained in this chapter.

34

Page 28: Chapter 2 - Literature Review 4 Dec

8