## 1. Introduction

Biometric systems are based on the use of certain distinctive human traits, be they behavioral, physicial, biological, physiological, psychological or any combination of them. As reflected in the literature, some of the most frequently used biometric modalities include fingerprint, face, hand geometry, iris, retina, signature, palm print, voice, ear, hand vein, body odor and DNA. While these traits may be used in an isolated manner by biometric recognition systems, experience has shown that results from biometric systems analyzing a single trait are often insufficiently reliable, precise and stable to meet specific performance demands (Ross et al. 2006). In order to move system performance closer to the level expected by the general public, therefore, novel biometric recognition systems have been designed to take advantaje from taking multiple traits into account.

Biometric fusion represents an attempt to take fuller advantage of the varied and diverse data obtainable from individuals. Just as is the case with human recognition activities in which decisions based on the opinions of multiple observers are superior to those made by only one, automatic recognition may also be expected to improve in both precision and accuracy when final decisions are made according to data obtained from multiple sources.

its discussion of data fusion in biometric systems, the present chapter will analyze distinct types of fusion, as well as particular aspects related to the normalization process directly preceding data fusion.

## 2. Biometrics

Biometric recognition involves the determination of the identity of an individual according to his/her personal qualities in opposition to the classical identification systems which depend on the users’ knowledge of a particular type of information (e.g., passwords) or possession of a particular type of object (e.g., ID cards).

In biometrics, ‘recognition’ may be used to refer to two distinct tasks. In the first one, that is called verification, an individual claims to be certain user who has been previously registered(enrolled) into the system. It is also possible that the individual does not indicate his/her identity but there exist some additional information that allow to suppose it. In this case system operation is reduced to confirm or reject that a biometric sample belongs to the claimed identity. In the second task, identification, it is not available such prior information about individual’s identity, the system must determine which among all of the enrollees the subject is, if any. In the present chapter, only verification will be discussed insofar as identification may be understood as the verification of multiple entities.

In both cases (verification and identification), a sample of a predetermined biometric trait (e.g., face, voice or fingerprint) is captured (e.g., via photo, recording or impression) from a subject under scrutiny (donor), this is done using an adequate sensor for the task (e.g., camera, microphone or reader/scanner). A sample is called genuine when its donor identity and the identity of the claimed user are the same and it is called impostor when they are not. Following its capture, the sample is processed (feature extraction) in order to obtain the values of certain predefined aspects of the sample. This set of values constitute the feature vector. The feature vector is then matched against the biometric model corresponding to the individual whose identity has being claimed[1] - . This model has been created at the time of that user enrols into the system. As a result of the matching, an evaluation of the degree of similarity between the biometric sample and the model is obtained, it is called score[2] - . With this information, the decision is taken either to accept the subject as a genuine user or to reject the subject as an impostor (see Fig. 1).

In the majority of biometric recognition systems currently in use only a single biometric trait is captured in order to confirm or reject the claimed users’s identity. Such systems are known as monobiometric. Nevertheless, at they heart is a pattern recognizer, which arrives at a final decision with the results obtained from a single sample processed according to a single algorithm.

Fig. 1 presents a simple representation of the biometric recognition process in monobiometric systems. After a subject presents the biometric trait which the system’s sensor is designed to process, in the first stage (i.e., capturing sample), a biometric sample is obtained by the sensor and processed by the system to eliminate noise, emphasize certain features of interest and, in general, prepare the sample for the following stage of the process. In the next step (i.e., feature extraction), characteristic parameters of the sample are quantified and a feature vector that groups them is obtained. Following quantification, the system proceeds to match the feature vector (i.e., model matching) against others captured during the training phase that correspond to the individual whose identity is being claimed. These latter vectors are often represented in biometric systems with models that summarizes their variability. As a result of the matching process, a score is obtained quantifying the similarity between the sample and the model. In the final stage (i.e., decision making) and as a result of the score generated, the biometric system makes a decision to accept the sample genuine or to reject it as impostor.

## 3. Biometric fusion

In polybiometric systems, various data sources are used and combined in order to arrive at the final decision about the donor’s identity. These systems are composed of a set of monobiometric parallel subprocesses that operate the data obtained from the distinct sources in order to finally combine them (i.e., fusing data). This fused data is then processed by system through a single subprocess until th final decision can be made regarding the truth of the claimed identity

In the construction of polybiometric recognitions systems, certain parameters must be set in response to the following questions:

What are the distinct sources of biometric data being analyzed?

At what point in the biometric process will the data be fused or, said another way, what intermediate data will be used for this fusion?

What algorithm is most adequate for bringing about a particular type of fusion?

The following sections of the present chapter will analyze these different aspects of system design.

## 4. Data sources

In order to respond to the first question of the previous paragraph regarding multiple data sources, the following specific questions must also be considered (Ross 2007).

How many sensors are to be utilized? In the construction of multi-sensor systems, different sensors (each with distinct performances) are used in order to capture multiple samples of a single biometric trait. In one example of this sort of polybiometric system, simultaneous photographs are captured of a subject’s face using both infrared and visible light cameras.

How many instances of the same biometric trait are to be captured? Human beings can present multiple versions of particular biometric traits (e.g., fingerprints for different fingers, hand geometry and veins for each hand and irises for each eye). As a result and with a schema similar to that of multi-sensor systems, multi-instance systems are designed to capture various instances of the same biometric trait.

How many times is an instance of a particular trait to be captured? Using a single sensor and a single instance of a particular trait, it is nevertheless possible to obtain distinct samples of that instance under different conditions (e.g., video images taken of a trait instance from different angles or voice recordings taken at different moments and with different speech content). These multi-sample systems may also be represented by a schema similar to that of multi-sensor systems.

How many different biometric traits are to be captured? Biometric recognition systems may be designed to analyze a single biometric trait (i.e., unimodal systems) or various traits (i.e., multimodal systems). The particularities of the latter type of system are represented by the schema below.

How many distinct feature extraction algorithms are to be utilized in the processing of the biometric samples? Multi-algorithm systems are designed to use various algorithms for the feature extraction from biometric samples. In this case, the use of different extraction algorithms may allow the system to emphasize different biometric features of interest (e.g., spectral or prosodic features of a voice sample) and produce different feature vectors for each.

Against how many types of patterns and using how many methods are the feature vectors to be matched? Multi-matching systems are biometric recognition systems that allow match the feature vectors against various types of models or/and. using multiple techniques.

Finally, it is also possible to construct hybrid systems systems of an even greater complexity that incorporate more than one type of the multiple data source discussed above.

## 5. Fusion level

As discussed earlier, biometric fusion is composed by a set of monobiometric subprocesses that work in parallel to operate the data obtained from distinct sources. Once this different data has been operated and fused, it is then handled by the system through a single subprocess until the point where the donor’s identity final decision can be made. This process is represented in Fig. 6 below.

Having considered the biometric fusion schema, it is time to return to the questions articulated earlier in the chapter and analyze now at what level of the process the fusion should be carried out or, in other words, what type of data the system should fuse. The possible responses in the literature to these points allow to establish diverse characterizations of data fusion systems defined as fusion levels(Ross 2007) (Joshi et al. 2009) (Kumar et al. 2010).

The first point at which data fusion may be carried out is at the sample level, that means immediately following sample capture by system sensors. This type of fusion is possible in multi-sensor, multi-instance and multi-sample systems and may be obtained by following a particular sample fusion method. The form that this method takes in each case depends on the type of biometric trait being utilized. While fusion may range from a simple concatenation of the digitalized sample data sequence to more complex operations between multiple sequences, but it is almost always carried out for the same reason: to eliminate as many negative effects as possible associated with the noise encrusted in the data samples during capture. Once the fused sample has been generated, it may be used by the system for feature extraction.

The second point at which data may be combined is immediately following the feature extraction. At the feature level, vectors derived from the different sources are combined, yielding a single, fused vector.

Another alternative is the fusion of scores obtained following the matching of different sample data against corresponding models. The new score resulting from this fusion is then used by the system to reach the final decision. This sort of fusion is normally carried out according to mathematic classification algorithms ranging in type from decision trees to techniques from the field of artificial intelligence, the latter of which offering learning capabilities and requiring prior training. The present chapter focuses particularly on this latter type of fusion which will be developed in much greater detail in sections below.

Fusion may also be carried out on the final decisions obtained for each monobiometric process through the use of some kind of Boolean function. The most frequent algorithms used in this type of fusion are AND, OR and VOTING. With the first type, the final, combined decision is *GENUINE* if and only if each monobiometric process decision is also *GENUINE*. For the second type, the final, combined decision is *IMPOSTOR* if and only if each monobiometric process decision is also *IMPOSTOR*. Finally, for the third type combined decision is that of the majority of monobiometric process decisions which may or may not have been previously weighted.

Finally, dynamic classifier selection schema uses scores generated at the data matching level in order to determine what classifier offers the highest degree of confidence. The system then arrives at a final decision through the application of solely the selected classifier. This is represented in Fig. 11 below.

## 6. Biometric performances

For the recognition of a individual by a classical recognition system, the data collected (e.g., passwords or ID cards information) from the subject must be identical to the previously recorded data into the system. In biometric recognition systems, however, almost never the data captured from a subject (nor the feature vectors obtained from them) are identical to the previous ones (Ross et al. 2006). The reasons for these variations are manifold and include the following:

Imperfections in the capture process that create alterations (e.g., noise) in the data;

Physical changes in the capture environment (e.g., changes in lighting and degradation of the sensors used); and

Inevitable changes over time in individual’s biometric traits.

As a result of the unrepeatibility of biometric samples, the process of biometric recognition can not be deterministic and it must be based on the stochastic behaviour of samples. In this way, rather than flatly asserting correspondence between a biometric sample and the corresponding model, biometric systems only permit the assertion that this correspondence has a certain probability of being true.

The differences observed among the distinct biometric samples taken of a single trait from a single donor are known as intra-class variations. On the other hand, inter-class variation refers to the difference existing between the samples captured by the system from one subject and those of others. The level of system confidence in the correctness of its final decision is determined according to these two types of variation. The lesser the intra-class variation and the greater the inter-class variation are, the greater the probability that the final decision is correct.

In the matching model step, the system assigns a score to the sample feature vector reflecting the system’s level of confidence in the correspondence between sample and claimed identity. If this score (*s*) lies above a certain threshold (*th*) (i.e., if: s ≥ th), the system will decide that the sample is genuine. However, if the score lies below the threshold the system will decide that the sample is an impostor one.

Insofar as score, as understood here, is a random variable, the probability that any particular score corresponds to a genuine sample can be defined by its probability density function (pdf) *f*
_{
g
}
*(s)*. Similarly, the probability that the score corresponds to an impostor sample can be defined by a pdf *f*
_{
i
}
*(s)*. As a result, the terms ‘false match rate’ (FMR) or ‘false acceptance rate’ (FAR) may be defined as the probability that an impostor sample be taken by the biometric system as genuine. Similarly, the terms ‘false not match rate’ (FNMR) or ‘false rejection rate’ (FRR) may be defined as the probability that a genuine sample be taken for as an impostor one.

When the decision score threshold is established in a system (see Fig. 12), the level of system performance is therefore established, because FAR and FRR directly depend on its value. Wether threshold values increases, FAR will also increase while FRR will decrease [Stan et al. 2009]. The optimal value of *th* can be obtained by minimizing the cost function established for the concrete system use. This cost function defines the balance between the damage that can be done by a false acceptance (e.g., a subject is granted access by the system to a protected space in which he or she was not authorized to enter) and that done by a false rejection (e.g., a subject with authorization to enter a space is nevertheless denied entry by the system).

The National Institute of Standards and Technology (NIST) proposes as a cost function the one shown in formula 2, which is a weighted sum of both error rates. C_{FR} and C_{FA} correspond to the estimated costs of a false rejection and false acceptance, respectively, and P_{g} and P_{i} indicate the probabilities that a sample is genuine or impostor. Is obviously true that *Pi+Pg=1* (Przybocki et al. 2006):

In NIST recognition system evaluations, the costs of a false acceptance and a false rejection are quantified, respectively, at 10 and 1, whereas the probabilities that a sample is genuine or impostor are considered to be 99% and 1%, respectively. With these parameters and normalizing the resulting expression, the following equation is obtained (formula 3):

For reasons of expediency, however, the present chapter utilizes other criteria that nevertheless enjoy wide use in the field. According to these criteria, C_{FA} = C_{FR} and P_{g} = P_{i}, such that the resulting cost function may be defined as the following (formula 5):

Another value used in the characterization of biometric systems is the equal error rate (EER) which, as shown below, indicates the point at which the error rates are equal:

As a final concept to consider here, the receiver operating characteristic curve (ROC curve) is a two-dimensional measure of classification performance and is regularly used in the comparison of two biometric systems. The ROC curve represents the evolution of the true acceptance rate (TAR) with respect to that of the FAR (Martin 1997):

Through the analysis of the ROC curve, the evaluation of a recognition system may be carried out by considering the costs associated with errors even where the latter have not been previously established. In particular, using the area under the convex ROC curve (AUC), system performance may be expressed as a single numeric value and evaluated: the system considered the best being that with the greatest AUC (Villegas et al. 2009)(Marzban 2004).

## 7. Single scores distribution

Let the simplest case of match score distribution be supposed where, for a single source, scores are distributed according to the following criteria:

Given the symmetry of the functions, it can be held that the threshold value minimizing the cost function can be located at *th=0*, point at which FAR and FRR are equal, defining also the EER as shown in formula 9:

From an estimation, the value *EER=15.85%* is obtained. It is clear, then, that the farther apart the centroids or the smaller the deviations of the distribution functions are, the smaller the error rates.

## 8. Multiple score fusion

Let it be supposed that match score fusion is to be applied to the results of two processes having generated independent scores (s_{1} and s_{2}) and with distribution functions identical to those described in the previous section of the present chapter. Thus, a match score vector is formed with Gaussian distribution functions for both genuine and impostor subject samples. This vector will have two components, each of which integrating the results from each of the monobiometric classifiers.

(10) |

In Fig.15, the distribution functions are presented together for both genuine and impostor subject score vectors. Right image represents the contour lines of the distribution functions. Observing it, it seems intuitive that, just as was done in the previous section of the present chapter and applying the criteria for symmetry discussed therein, the best decision strategy is that which takes as a genuine subject score vector any vector found above and to the right of the dotted line which, in this particular case, corresponds to *s1+s2 ≥ 0*. This converts the threshold, for the one-dimension scores, to a boundary line decision in this two-dimension space (an hiperplane if n dimensions space).

Following this, the resulting estimation of the EER is shown in formula 12. In the specific case proposed here, the resulting EER is found to be 7.56% indicating an important improvement owing to the fact that the centroids of the distribution functions have been separated here by a factor of

## 9. Using gaussian mixture model classifiers

Gaussian mixture model (GMM) classifiers are used in order to create a model of statistical behaviour represented by the weighted sum of the gaussian distributions estimated for the class of genuine training score vectors and another similar model to represent the class of impostor vectors. Using the two models, the vectors are classified using the quotient of the probabilities of belonging to each of the two classes. If this quotient is greater than a given threshold (established during the system training phase), the vector is classified as genuine. If the quotient is below the given threshold, the vector is classified as an impostor. Such a procedure is quite similar to that discussed in the previous section of the chapter.

In a situation such as that described in the paragraph above, the following points indicate the expectations for a training process and test using GMMs:

These models (f

_{g}’, f_{i}’) of sums of Gaussian functions should maintain a certain similarity to the generative sample distribution ;The established threshold may be equivalent to the theoretical decision boundary between genuine and impostor score vectors; and

Test results clearly approach the theoretic FAR and FRR.

In order to test the fitness of these premises, 1000 two-dimensional random vectors (Vg) following the distribution function of the genuine vectors and another 1000 vectors (Vi) following the distribution function of the impostor vectors have been taken as training data. With these vectors, GMMs were created to approximate the distribution functions[4] - .

The models obtained in the training phase for 10 Gaussian models (10G) derived from the simulated data training are presented below in Table 1:

In Fig.16 (left), genuine and impostor models are presented for the score s_{1} of the score vector. With red lines indicating the impostor model and black lines indicating the genuine sample model, each of the 10 individual Gaussian distributions with which the GMM classifier approximated the distribution of the training data are represented by the thin lines on the graph. The weighted sums of these Gaussian functions (see Formula 13) are represented by the thick lines on the graph. The result has an appearance similar to two Gaussian distributions around +1 and -1. Fig. 18 (right) shows the contour lines of the two-dimensional models.

For a value of *th = 0.9045* (calculated to minimize) it was found that FAR = 8.22% and FRR = 7.46%.

In Fig.16. the decision boundary line, at which the quotient of pdfs is equal to the threshold and which separates genuine and impostor decisions, presented as a dotted line. This line is quite near to the proposed boundary. Then the formula 14 represents a transformation from a two-dimension criterion to a one-dimension threshold, which, of course, is easier to manage.

If the same exercise is repeated for a model with 3 Gaussians (3G) and for another with only 1 Gaussian (1G), the following results are obtained:

N Gaussian | FAR | FRR | MER^{}
| th | AUC |

10 | 8.22% | 7.46% | 7.84% | 0.9045 | 97.12% |

3 | 7.97% | 7.52 % | 7.75% | 0.9627 | 97.05% |

1 | 7.99% | 7.56% | 7.77% | 0.9907 | 97.10% |

Changing the threshold value (see Fig.19), distinct decision boundaries and their corresponding error rates may be obtained. With these values, a ROC curve may be drawn and its AUC estimated.

## 10. Using support vector machine classifiers

A support vector machine (SVM) is a classifier that estimates a decision boundary between two vector sets (genuine and impostor ones) such that maximizes the classification margin. In the training phase of a SVM, a model is constructed that defines this boundary in terms of a subset of data known as support vectors (*SV*), a set of weights (*w*) and an offset (*b*).

The equation above defines the distance of a vector (v) to the boundary, where positive distances indicate genuine samples and negative distances indicate impostor samples[6] -
. For other kind of boundary lines is possible to select between different kernel functions. Then the general decision function is shown in formula 16, where *K(sv,v)* represents the adequate kernel function. The kernel implied in formula 15 is called “linear kernel”.

Given the data distribution and the fact that the expected separation boundary is a straight line, it may be assumed that the linear kernel is the most adequate kernel function here.

Fig.20 shows the distribution of genuine samples (in blue) and impostor samples (in red). Points indicated with circles correspond to the support vectors generated in the training phase. The central black line crossing the figure diagonally represents the set of points along the boundary line, which is also quite close to the theoretical boundary.

The results of the test data classification demonstrate the performance indicated below for

The classifier establishes a transformation of the vector space into a real value whose module is the distance from the boundary, calculated such that the system be optimized to establish the decision threshold at the distance of 0. Just as in the case of GMMs, system behavior can be analyzed using the ROC curve and, more specifically, the AUC through the adjustment of this threshold value (see Table 3).

## 11. Using neural network classifiers

An artificial neural network (ANN) simulates an interconnected group of artificial neurons using a computational model. In this context, a neuron is a computational element that operates n-inputs in order to obtain just one output following a transfer function like the one shown at formula 16. Where s_{
k
} is the k-esime neuron input, w_{k} is the weigth of k-esime input w_{0} represent the offset and finaly represents a function (typically sigmoid or tanh) that performs the transference between neurons.

A typical ANN groups its neurons in a few layers, so that, the certain layer neuron outputs are only connected to the next layer neuron inputs.

The neural network training step gets as a result the weight for every neuron input that minimizes the error rates.

Then the simplest network is one which has only one neuron with two input and one output (2-1-1). This way, the transfer function has no effect on the system and at the end decision function becomes a linear combination of the inputs and therefore the training estimates a linear separator similar to the one seem before for SVM with linear kernel.

Applying neural networks to above described data, is possible to obtain the following results:[7] -

## 12. Beta distributions

One common way in which monobiometric systems present their scores is through likelihood estimates (the probability that the sample is genuine). In such cases, the score rangeis limited to 0-1 (0-100%). Ideally, instances of genuine subject scores would be grouped together around 1 or a point close to 1, while impostor subject scores would be grouped together around zero or near it. Both would demonstrate beta distributions. An example of this ideal situation is plotted in Fig.23 with the pdf for genuine samples follows *Beta(5,1)* and the pdf for impostor samples follows *Beta(1,5)*. Because the symmetrical properties of these functions, the equilibrium point can be clearly located at *s = 0.5* with an solving the integral in formula 1 the *EER = 3.12%*.

As it was done for the Gaussians, identical distribution functions are established for both dimensions of the two-dimension score space, then a theoretical value of EER= 0.396% would be obtained. Also is possible the same routine and evaluate system performances for GMM, SVM and NN classifiers[8] - .

Fig. 21 shows the pdf’s used in this example and de model obtained for them, while table 5 display test results.

## 13. More realistic distributions

Unfortunately, the distribution functions for real scores are not as clear-cut as those presented in Fig. 21. Scores for impostor subject samples, for example, are not grouped around 0, but rather approach 1. Similarly, genuine subject sample scores often tend to diverge from 1. Distributions similar to those in Fig. 22 are relatively common. To illustrate this, *pdf.genuine = Beta (9,2)* and *pdf.impostor = Beta (6,5)* have been chosen for the first score (s_{1}).

These particular distributions don’t display any symmetrical property then the equilibrium point estimated loking for *FAR = FRR* and as a result the threshold value of *th = 0.7* with an *EER of 15.0%* has been obtained. Eve more, the optimal threshold value does not coincide here with the ERR. Then in order to minimize the cost function threshold must adopt a value of *0.707* yielding the error rates of *FAR = 16.01%, FRR = 13.89%* and *MER = 14.95%*.

In order to further simulate real conditions, score 2 has been supposed here to display a different behavior, to wit, *pdf.genuine = Beta (8,4)* and *pdf.impostor = Beta (4,4)*, as it is displayed in Fig 22

As can be seen, the equilibrium point is found here at a value of *EER = 29.31%* and *th = 0.5896* and the minimum of the cost function at *FAR =34.84%, FRR = 23.13%, MER = 28.99%* and *th = 0.5706%*.

If these two distributions are combined and a two-dimensional score space is established, the resulting pdfs can be represented as the one in Fig.23. It plots these two-dimensional density distributions where de genuine one is found near the point (1,1) while the impostor one is located farther from it.

Applying the GMM trainer with 10 Gaussian functions to these distributions, the images in Fig. 24 are obtained representing the set of the 10 Gaussians making up the genuine model; the set of 10 Gaussians making up the impostor model and representing the impostor and genuine models as the weighted sum of each of their Gaussian functions.

Equivalent representations can be obtained using a GMM with 3 Gaussians

As in previous sections tests conducted with GMM, SVM and ANN classifiers yield the following results:

## 14. Match score normalization

As described in earlier sections of the present chapter, the data sources in a system of match score fusion are the result of different monobiometric recognition subprocesses working in parallel. For this reason, the scores yielded are often not homogeneous.

In the most trivial case, the source of this lack is different meaning of the scores, they may represent the degree of similarity between the sample and the model or the degree of disimilarity or directly represent the degree of subsystem confidence in the decision made.

Other sources of non-homogeneous scores are include the different numeric scales or the different value ranges according to which results are delivered, as well as the various ways in which the non-linearity of biometric features is presented. Finally, the different statistical behavior of scores must also be taken into account when performing the fusion. For these reasons, the score normalization, transferring them to a common domain, is essential prior to their fusion. In this way, score normalization must be seen as a vital phase in the design of a combination schema for score level fusion.

Score normalization may be understood as the change in scale, location and linearity of scores obtained by distinct monobiometric recognition subprocesses. In a good normalization schema, estimates of transformation parameters must not be overly sensitive to the presence of outliers (robustness) and must also obtain close to optimal results (efficiency) (Nandakumar et al. 2005)(Jain et al 2005)(Huber 1981)

There are multiple techniques that can be used for score normalization. Techniques such as min-max, z-score, median and MAD, double sigmoid and double linear transformations have been evaluated in diverse publications (Snelick et al. 2003) (Puente et al. 2010).

## 15. Min-max normalization

Perhaps the simplest of currently existing score normalization techniques is min-max normalization. In min-max normalization, the goal is to reduce dynamic score ranges to a known one (tipically: 0-1) while, at the same time, retaining the form of the original distributions.

For the use this technique, it is necessary that maximum and minimum values (*max*, *min*) were provided by the matcher prior to normalize. Alternatively, these may be evaluated as the maximum and minimum data of the values used during system training. In this way, a linear transformation is carried out where 0 is assigned to the minimum value 1 to the maximum value[9] -
. This transformation function is shown in the following formula 16:

Applying this transformation to the observations generated in a previous section of the current chapter (see ’13. More realistic distributions’), the following results are obtained:

## 16. Z-score normalization

Due to its conceptual simplicity, one of the most frequently used transformations is z-score normalization. In z-score normalization, the statistical behavior of the match scores is homogenized through their transformation into other scores with a mean of 0 and a standard deviation of 1.

Clearly, it is necessary that the mean and standard deviation of the original match scores be known prior to normalization or, as in min-max normalization, they should be estimated from training data.

Z-score distributions do not retain the forms of the input distributions, save in cases of scores with a Gaussian distribution, and this technique does not guarantee a common numerical range for the normalized scores.

Test results from z-score normalization are shown below:

## 17. Median and MAD

The median and MAD (median absolute deviation) normalization technique uses the statistical robustness resulting from the median of a random distribution to make it less sensitive to the presence of outliers. Nevertheless, median and MAD is generally less effective than z-score normalization, does not preserve the original distribution and does not guarantee a common range of normalized match scores.

## 18. Double sigmoid normalization

In one particular study from the literature, a double sigmoid transformation is proposed as a normalization scheme (Cappelli et al. 2000):

According to the double sigmoid normalization technique, match scores are converted to the interval [0,1]. While the conversion is not linear, scores located on the overlap are nevertheless mapped onto a linear distribution (Fahmy et al. 2008).

## 19. Double linear normalization

Scores yielded by monobiometric classifiers are interpreted as pair of a decision and confidence. The decision, thus, is made according to the location side of the score is located respect to the threshold while confidence is de distance between them. Thus, the greater the distance to the threshold, the greater will be the weight assigned to the score for the final decision.

Generally, this distance does not enjoy a homogeneous distribution for scores of genuine and impostor observations. As a result, in distributions such as that presented in Fig.23, scores of impostor samples tend to have a greater likelihood than those of genuine scores.

In order to compensate that kind of heterogeneity, a transformation has been proposed in (Puente et al. 2010) to make distributions more uniform around the decision threshold:

## 20. Conclusions

The principal conclusion that can be drawn from the present chapter is undoubtedly the great advantage provided by score fusion relative to monobiometric systems. In combining data from diverse sources, error rates (EER, FAR and FRR) can be greatly reduced and system stability greatly increased through a higher AUC.

This improvement has been observed with each of the classifiers discussed in the present chapter. Nevertheless and in consideration of comparative studies of normalization techniques and fusion algorithms, it can be noted that the specific improvement produced depends on the algorithms used and the specific case at hand. It is not possible, therefore, to state *a priori* which techniques will be optimal in any given case. Rather, it is necessary to first test different techniques in order to pinpoint the normalization and fusion methods to be used.

One final conclusion that stands out is that improvements in error rates are directly linked to the number of biometric features being combined. From this, it may be deduced that the greater the number of features being fused, the larger the improvement will be in the error rates.