Differential gradient operators.

## Abstract

The network of blood vessels possesses several properties that make a good biometric feature for personal identification: (1) they are difficult to damage and modify; (2) they are difficult to simulate using a fake template; and (3) vein information can represent the liveness of the person. In the process of recognition of the network of blood vessels, we encounter two main difficulties: the first difficulty concerns the enhancement of the image of blood vessels obtained from the camera working in visible and/or infrared light, and the second one concerns the process of extraction of features and methods of classification. In the first part, this chapter presents the basic methods of preprocessing biometric images. In the second part, we discuss the process of feature extraction with particular emphasis on the feature extraction from images depicting the network of blood vessels. This applies to texture analysis using the co-occurrence matrix, Gabor filtration, moments, and topological features using cross points. In the third part, we present the methods of processing images of the blood vessel network of dorsal part of the hand and wrist. We also discuss the process of reducing the dimensionality of a feature vector using the principal components analysis method.

### Keywords

- biometrics
- vein patterns
- feature extraction
- co-occurrence matrix
- Gabor’s filters
- classification

## 1. Introduction

Biometrics is a powerful field of science for identifying a person using their physiological and behavioral features [1, 2]. Biometrics is the automatic recognition of people based on behavioral or physiological characteristics. During recognition given users are assigned to prescribed classes. We extract the essential features of the object and use these features to classify the object.

Biometric systems in general perform two tasks: identification and verification (recognition) of people (Figure 1). The process of verification (recognition) boils down to distinguishing a specific person from a limited number of people whose biometric data are known. The identification consists of determining the vector of features corresponding to the person being subjected to the identification process and trying to find a match between this vector and the feature vectors in the database containing records (feature vectors) concerning people. As a result, we get a list of the most similar individuals in the database. Identification is much more difficult [3, 4].

Images play an important role in the identification process of people. Image processing and recognition are fields that use complex signal and image processing algorithms.

The image in digital form is stored as a two-dimensional array. Formally

and

where *G-1* is the gray/color maximum value of each resolution cell.

The components of an image processing system are presented on Figure 2.

The processing generally comprises the steps of acquiring an image, selecting the desired color space, improving image quality, image segmentation, and features extraction for the recognition. Recognition process involves several stages—extraction features and dimensionality reduction which selects the best set of features and rejects irrelevance. The resultant feature vector is the basis for classification.

The image is usually obtained using a CCD camera or NIR camera. It can be a color image (three-color components) or a grayscale image. Usually, color space (RGB with 24 bit) is converted to gray color space (8 bit).

Below, some steps shown in image processing system on Figure 2 will be explained in more detail [5].

Image processing operations can be divided into (Figure 3):

Processing of single points of the image.

Operations that use pixel group processing.

The first group includes operations related to modification histogram, while the second group includes operations related to edge detection and various types of image filtration.

Transforming the brightness scale of image elements enables:

In the case where the brightness range does not cover the entire scale available for the image, the extension of the range (the effect of increase contrast)

Emphasizing certain brightness ranges and suppressing others

Modifying the brightness of image elements to obtain a uniform image frequency of the occurrence of appropriate levels of brightness

In practice, transformation *T* can be a logarithmic transformation, exponential transformation, etc. (Figure 4).

If *g*, e.g.,

The gray levels are modified as [5, 6]

where

One of the methods of noise elimination (“salt pepper” type) and other image distortions is median filtering (MF). Median filtering is a nonlinear operation, and this fact complicates the mathematical analysis of its properties. It is implemented by moving the window (the mask) along the lines of the digital image and changing the value of the middle window element by the median value of the elements inside the window. MF allows you to keep sharp changes in brightness and high efficiency in eliminating impulsive noise [5].

The 2D *MF* for an image

where *MF* window.

MF allows you to keep sharp changes in brightness and high efficiency in eliminating impulsive noise (Figure 7).

Edges carry useful information about object boundaries which can be used for further analysis. Edge detectors can be grouped into two classes: (a) local techniques which use operators on local image neighborhoods and (b) global techniques.

Gradient estimates is done as

and can be expressed by (Table 1)

or

Edge detector operators | Partial derivatives along x and y axes | Weight vectors | Kernels |
---|---|---|---|

Differential | |||

Roberts edge detectors | |||

Max. difference | there is no | ||

Prewitt edge detector | |||

Sobel edge detector |

where

Another popular operator, not shown in Table 1, is the Canny edge detector operator implemented in accordance with the Figure 8 [7].

Examples of applications of edge detection operators are shown in Figure 9.

Let *X* a finite subset of the plane on which the function *X* into *K* non− empty subsets *Reg* the rule specified on the set *S* and assuming the value *true* if and only if any pair of points from each subset

The segmentation of the image *Reg* rule is the division

The *Reg* rule specifies a certain homogeneity criterion and depends on the function of

## 2. Feature extraction

Methods for feature extraction on biometric traits can be categorized into geometrical analysis and textural analysis (Table 2).

Biometric physiological modality | Geometrical features | Texture features |
---|---|---|

Fingerprint | Minutiae singular points Delta points Triangulation methods Crossing number | Analysis texture pattern composed with ridges and valleys Spatial distribution of minutiae points |

Palmprint | Principal lines. Line edge map Wrinkles Palmar friction ridges Shape-oriented features | Local line binary pattern Co-occurrence matrix |

Finger knuckle print | Shape-oriented features: lines, curves, contours | Curvelet Co-occurrence matrix Wavelets |

Hand geometry | Shape-oriented features Finger length and width | — |

Face | Spatial relationship among eyes, lips, nose, chin | Gabor’s filtering LBP |

Ear | Force field transformation 2D and 3D shape descriptors | Moment invariants |

Iris | — | Phase-based method Gabor’s filtering |

Periocular | Geometry of eyelids, eye folds, eye corners | LBP Histogram of oriented gradients SIFT (shift-invariant feature transform) |

Retina | Minutiae singular points Crossing number | Gabor’s filtering |

Vein Hand vein Finger vein Forearm vein | Bifurcation points Ending points | Gabor’s filtering Riesz transform Wavelet, curvelet Radon transform |

The texture image can be seen as an image area containing repetitive pixel intensity patterns arranged in a certain structural manner. The concept of texture has no formal and mathematical definition, but there are a number of methods for extracting texture features that can be roughly divided into model-based (fractal and stochastic method), statistical, and using signal processing algorithms.

Methods using signal processing algorithms (in the frequency domain and/or space-frequency domain) are widely used in transform-based texture analysis, e.g., Fourier transform, Gabor transform, Riesz transform, Radon transform, and wavelet transform.

One of the popular representations of texture feature is the co-occurrence matrix proposed by Haralick et al. [8, 9, 10]. The gray-level co-occurrence matrix (GLCM) *k* and *l* at a given distance *d* and then extracts statistical measures from this matrix. The element of co-occurrence matrix is defined as

These features provide information about the texture and are as follows:

The distance *d* is most often represented in polar coordinates in the form of a discrete distance and an orientation angle. In practice, we use four angles, namely,

Mathematically, Gabor filters is defined as [11]

Typically, Gabor’s filter bank was created by varying the frequency parameter, the orientation parameter, and the variance parameter (Figure 11).

Gabor’s features are obtained by convolution of the image *f (x, y)* with the

where

Moment-based features can be successfully used as elements of a feature vector in biometrics using blood vessel network [13, 14].

The geometric moments of order

where *x* is the degree *p*, while *y* is the degree *q*. If

Infinite set of moments

Central moments are defined by

where

Standardized central moments receiving as

where

We usually use the first seven combinations of central moments of order 3 known in the literature as Hu moments [15].

The basic set of geometrical moments is non-orthogonal which makes selection of features difficult.

Zernike’s moments are orthogonal and invariant to rotation, translation, and scale change. The complex set of Zernike’s moments is determined by [16]

where

When calculating Zernike’s moments, the size of the image determines the disk size, and the disk center is taken as the origin. In the case of considering moments on the order of 7, we get 20 Zernike’s moments.

In the case of biometric data using images of retinal blood vessels and conjunctival blood vessels, one of the stages of creating a vector of features is to determine geometrical features based on the topological properties of the image [5, 17].

The number of connected points around the point

where 4 denote the four-element neighborhood of the image point, *S* denotes the set of integers [17]. In the case where

If

The feature vector defining the topology of blood vessels is made up of the number of bifurcation points, number of crossing points, coordinates of bifurcation points, and coordinates of crossing points.

By using the relationship between the characteristic points of the user blood vessel image and blood vessel image of template, we can calculate the matching score results.

## 3. Vein biometrics: feature extraction from hand dorsal and wrist images

One of the most promising and intensively developed biometric methods is the method using the network of blood vessels. The pattern of blood vessels is unique for every human being and also in the case of twins. It is also stable over time [18]. Biometrics associated with the network of blood vessels has a significant advantage over other biometric methods, namely [1, 4, 18]:

Allows only identification of living people: the NIR camera records the image only in the case of deoxygenated hemoglobin, and this is possible only in the living organism [19, 26];

The network of blood vessels is inside the body, and it is practically impossible to reproduce outside of it, which results in very high level of safety.

Usually, we use the network of blood vessels associated with the following parts of the body:

Eye. This applies first of all not only to the retinal blood vessels but also to the blood vessels of the conjunctiva.

Hand. In this case, we are talking about the network of blood vessels of the finger, palm, hand dorsal, wrist, and forearm [20, 24, 25].

Figure 12 Shows the networks of blood vessels used in biometry.

We will consider images from Figure 12(e) and (f), which can be obtained in one process of acquiring biometric patterns. In the literature on the subject, the analysis of this type of images for biometrics is referred to as *dorsal vein biometrics* and *wrist vein biometrics* [27, 28].

### 3.1. Vein biometrics

In the process of identifying people on the basis of dorsal vein images, we use a feature vector constructed from two parts: features calculated on the basis of the co-occurrence matrix and features calculated using Gabor filtration operation [21, 22, 23].

We consider the dorsal vein images shown in Figure 13.

We analyze the co-occurrence matrix for *d* are shown in Table 3.

Figure 13a | Figure 13b | |||||||||
---|---|---|---|---|---|---|---|---|---|---|

IDM | Contrast | Energy | Entropy | Corr. | IDM | Contrast | Energy | Entropy | Corr. | |

0.210 | 50.890 | 3.218E-4 | 8.368 | 2.271E-4 | 0.255 | 141.318 | 0.007 | 7.662 | 2.926E-4 | |

0.173 | 67.564 | 2.622E-4 | 8.507 | 2.289E-4 | 0.218 | 198.051 | 0.006 | 7.794 | 2.932E-4 | |

0.244 | 31.936 | 3.479E-4 | 8.218 | 2.268E-4 | 0.306 | 86.913 | 0.007 | 7.464 | 2.934E-4 | |

0.146 | 80.621 | 2.375E-4 | 8.608 | 2.286E-4 | 0.214 | 199.180 | 0.006 | 7.803 | 2.931E-4 | |

0,111 | 179.516 | 1.944E-4 | 8.804 | 2.265E-4 | 0.184 | 378.481 | 0.005 | 7.991 | 2.928E-4 | |

0.102 | 206.593 | 1.757E-4 | 8.861 | 2.302E-4 | 0.161 | 456.028 | 0.004 | 8.065 | 3.010E-4 | |

0.155 | 88.528 | 2.386E-4 | 8.581 | 2.273E-4 | 0.226 | 198.554 | 0.006 | 7.772 | 2.963E-4 | |

0.078 | 288.156 | 1.626E-4 | 8.932 | 2.286E-4 | 0.161 | 459.783 | 0.004 | 8.075 | 3.002E-4 |

As a result, on the basis of the co-occurrence matrix, we obtain 40 features.

The second part of the feature vector is obtained by implementing an input image convolution operation with the bank of Gabor filters.

For each of the image, a filtration operation is carried out in accordance with Eq. (13) (Figures 14–16).

In the case of biometric identification of people based on texture features obtained using Gabor filter bank, we must solve the problem of a very large dimension of Gabor vector of traits.

### 3.2. Reduction of dimension of the feature vector by the PCA method

In the case of the 128 × 128 image and 3 × 6 of Gabor’s filter bank, the feature vector has a dimension of 128 × 128 × 3 × 6 = 294,912. The size of the feature is very correlated with each other; after down-sampling (according to factor 8), we get a vector of 36,864 elements or 2304 elements per image.

In order to reduce information redundancy, we use the principal component analysis (PCA) method. In some studies it is also called a Karhunen-Loeve discrete transform [29, 30].

The principal component analysis (PCA) method reduces the amount of data analyzed by subjecting them to linear transformation to a new coordinate system, resulting in new independent variables called the principal components.

The principal features of the PCA method are represented by eigenvectors. The eigenvectors of the covariance matrix are calculated based on the image training set and represent the principal components of the training image set.

We use two author’s database of images of the blood vessel network, namely, a database of dorsal vein images and a database of wrist vein images containing 42 images created as part of a session with students and 58 images found in the resources of www. Each database had 100 images.

The collection of training images consisted of 50 images (50% of images from the student base and 50% of images from web sources).

The PCA algorithm is made as follows:

Learning/training phase

The image *M × N* pixels and is converted into a 1 *× MN* size vector. Images from the training set are presented in the form of a *T* matrix (Figure 17):

where *q* is the number of images in the training set.

We calculate the mean image of all the images from the training database:

Then, we calculate the difference between each image from the training database and the mean image:

The covariance matrix is defined as

where

and matrix *A* has a dimension of *MN × q.*

The covariance matrix has a dimension of *MN × MN*.

Then, we calculate the eigenvalues and eigenvectors of the covariance matrix:

Then, we organize our eigenvectors according to their decreasing eigenvalues. We choose *k* principal components corresponding to *k* largest eigenvalues.

Test/recognition phase

The new image is processed to obtain eigenvectors and eigenvalues. *k* the main components of the

where

Approximated image is calculated as

We choose the *k* value according to the dependence:

where k is the predefined number of eigenvectors and *q* the total number of eigenvectors.

A high value of *k* means that a large amount of input information will be stored, e.g.,

The variance of the first eigenvector is about 60% of the variance of the data set, the variance of the first 30 eigenvectors is about 85% of the variance of the data set, and 45 or more eigenvectors account for over 90% of the variance of the data set (Figure 18).

By increasing the number of eigenvectors, we increase the recognition efficiency.

We defined the vectors of features as follows:

where

The quality of biometric systems is measured by two parameters: false acceptance rate *(FAR)* and false reject rate *(FRR). FAR* indicates the situation when the biometric input image is incorrectly accepted, and the *FRR* indicates the rejection of the user who should be correctly verified.

The size of the *FV1* vector has been set to 60 eigenvectors. The *featVect* size is 100. For these parameters *FRR* is 1.16% and *FAR* is 0.26%.

## 4. Conclusion

Recognition of people in biometric systems is based on the physiological or behavioral features that a person possesses.

In this chapter, we presented the image preprocessing operations used in static biometric systems (physiological modality). In particular, we discussed operations related to the transformation of the brightness scale of the image, modification of the brightness histogram, median filtering, edge detection, and image segmentation. In the course of these operations, we obtain an image enabling the extraction and measurement of features that serve as the basis for recognition.

Next, we discuss the feature extraction process, focusing on certain geometrical features and texture features. We present a representation of texture features based on parameters obtained from the co-occurrence matrix and images after Gabor’s filtration with various scaling and orientation parameters. In terms of geometric features, we discuss the moment-based features and geometrical features based on the topological properties of the image.

We provide and discuss the feature extraction process in the images of the blood vessels of the hand dorsal and wrist. We present features calculated on the basis of matrix of co-occurrences and texture characteristics obtained using Gabor’s filter bank. The process of reducing the dimensionality of a feature vector using the PCA method is also considered.

The main contributions of this chapter are the following:

A set of information on image processing methods used in biometric systems

Presentation of methods for obtaining feature vectors that are the basis of the process of recognizing people

Explaining the problem of reducing the dimensionality of the feature vectors

Showing parameters that are the basis for recognizing people based on images of the blood vessels of the hand dorsal and wrist

The chapter can be the basis for further studies and works on the image processing in biometric systems, especially based on images of the blood vessel network.