This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\history

Palm Vein Recognition via Multi-task Loss Function and Attention Layer

JIASHU LOU1    JIE ZOU2, AND BAOHUA WANG.3 College of Mathematics and Statistics, Shenzhen University, No. 3688, Nanhai Avenue, Shenzhen, Guangdong, China [email protected] [email protected] [email protected]
Abstract

With the improvement of arithmetic power and algorithm accuracy of personal devices, biological features are increasingly widely used in personal identification, and palm vein recognition has rich extractable features and has been widely studied in recent years. However, traditional recognition methods are poorly robust and susceptible to environmental influences such as reflections and noise. In this paper, a convolutional neural network based on VGG-16 transfer learning fused attention mechanism is used as the feature extraction network on the infrared palm vein dataset. The palm vein classification task is first trained using palmprint classification methods, followed by matching using a similarity function, in which we propose the multi-task loss function to improve the accuracy of the matching task. In order to verify the robustness of the model, some experiments were carried out on datasets from different sources. Then, we used K-means clustering to determine the adaptive matching threshold and finally achieved an accuracy rate of 98.89% on prediction set. At the same time, the matching is with high efficiency which takes an average of 0.13 seconds per palm vein pair, and that means our method can be adopted in practice.

Index Terms:
Palm vein recognition, Deep learning, Transfer learning, Multi-tasks loss function
\titlepgskip

=-15pt

I Introduction

In today’s online information society, we increasingly need to prove our identity in different public places. Previously, we used traditional methods of identification such as passwords, ID cards, family registers, and driving licenses, but these methods are easy to forget, lose and steal, and they have a number of security risks. In the era of Big Data, we have high-performance computers that have the computing power to take and process unique biological characteristics as a ”password” for identification, such as the face, fingerprint, palm print, iris, voice, and other biometric characteristics.

Among these, palmprint and palm vein recognition has become very effective technology in the field of biometric identification. In contrast to other biometric features, there are many available lines in the palm area, such as main lines, mastoid lines, and wrinkles. At the same time, palm veins have the advantages of high differentiation, stability, and low invasiveness. Currently, palm print and palm vein recognition has received a great deal of attention from researchers and is widely used in criminal investigation and public security.

In this paper, we propose a convolutional neural network based on VGG-16 transfer learning fused attention mechanism which is used as the feature extraction network on the infrared palm vein dataset. In the training phase, we consider the recognition of palm veins as a classification task, and in the prediction we need to match the palm veins two by two. Therefore, we introduce the multi-task loss function, which is found to be effective in improving the matching accuracy after experiments. Then, we use clustering to determine the adaptive matching threshold and finally achieve an accuracy rate of 98.89% on prediction set.

II Related work

II-A Palm Vein Recognition

Biometric identification technology uses an individual’s unique biometric characteristics to uniquely determine an individual’s identity [1][2]. With the development of mobile Internet and the increase of computing speed of personal electronic devices, technologies such as facial recognition [5][6], fingerprint recognition [3][4], and even iris recognition [7] have been integrated into every personal terminal. Moreover, palmprint recognition as an emerging biometric technology has recently appeared in the public eye. Palmprints have richer features than fingerprints, which are not easy to change and have better identification and uniqueness.

Palmprint research uses high-resolution or low-resolution images. High-resolution images are suitable for forensic applications such as criminal investigation, e.g., Jain, AK (2009)[11] developed latent palmprint matching, mainly dealing with low-resolution palmprints. Since about 30% of the biopsies recovered from crime scenes come from the palm, the evidentiary value of palm prints in forensic applications is evident. On the other hand, low-resolution images are more suitable for civilian and commercial applications such as access control. Generally, high resolution is defined as 400 dpi or higher, and low-resolution is defined as 150 dpi or lower. Ridges, singularities, and detail points are generally extracted as features in high-resolution images. While in low-resolution images, main lines, wrinkles, and textures are usually extracted. Initially, palmprint research focused on high-resolution images [29][30]. However, recent research hotspots have shifted to commercial and residential low-resolution palmprint recognition and matching [31]

Palm vein features are similar to it, and its application is better [38] because the features of veins are more evident and unique. Moreover, compared to palm prints, palm veins rely on biometric information inside the body and are therefore less susceptible to destruction, alteration, or tampering. Therefore, vein recognition is becoming one of the most reliable methods in biometrics and has attracted broad interest from biometric researchers. Veins are vast networks of blood vessels under the human skin that are almost invisible to the human eye and are more difficult to replicate than other biometric features [40]. The shape of the vascular pattern is thought to be unique across individuals (Wilson, 2010)[39] and stable over time.

For now, the academic community mainly focuses on the extraction and localization of palmprint features using mathematical means, such as Huang (2008) et al. [9] and Wu (2006) et al. [10] using intrinsic features of palmprints, such as main lines and wrinkles, for palmprint recognition. ; Lu, GM (2003) [8] proposed a method to extract palmprint feature vectors by Karhunen-Loeve transformation; Zhang, D (2010) extended the feature space to three dimensions, extracted depth features using structured light techniques, and developed a multilevel framework for personal identity verification. Palmprint matching has been richly used in various fields. In addition, deep learning networks are gradually used in palmprint recognition with the popularity of deep learning and the increase of computing hardware arithmetic power. For example, Zhao, SP (2022) [12] used deep convolutional neural networks to extract palmprint features and developed a joint constrained least squares regression framework for palmprint recognition; Trabelsi, S (2022) [13] used a simplified PalmNet-Gabor method to improve PalmNet and obtained a higher high accuracy, reducing the number of features and saving computation time.

II-B Convolutional Neural Networks

Convolutional neural networks (CNNs) use the original image as input, which can effectively learn the corresponding features from many samples and avoid the complicated feature extraction process. Unlike fully connected neural networks, CNNs can directly process two-dimensional images, while the former loses dimensional information in image spreading. Therefore, CNNs have been widely used in image processing. For example, R Vinoth (2014)[27] et al. used CNN networks for tumor image recognition and segmentation in magnetic resonance imaging and compared them with traditional methods such as SVM. Xiao Huang (2019)[28] et al. used two CNNs to extract visual and textual features from social media posts and identify disasters associated with related social media for fast response. It can be seen that convolutional neural networks have made outstanding contributions in the field of image recognition.

However, a deep convolutional neural network contains a large number of parameters to be trained (e.g., ResNet-50[17] has 23 million parameters), and we need a large amount of labeled data to train it, but manually labeling the data would be a tremendous amount of work. One idea is to use unsupervised learning to annotate the data, as in Kiselev, VY (2019)[15] Unsupervised learning of single-cell RNA-seq data using clustering. Another direction is transfer learning [16], which can take knowledge from one domain (source domain) and migrate it to another domain (target domain), enabling the target domain to achieve better learning results. This is especially true for scenarios where the source domain has sufficient data, and the target domain has a small amount of data. In image recognition and classification, transfer learning has achieved good performance. Shin, HC (2016)[14] used a network pre-trained on ImageNet dataset for medical image recognition by transfer learning for thoracoabdominal lymph node (LN) detection and interstitial lung disease (ILD) classification and reported achieving the best performance. The combination of CNNs and transfer learning allows for a much broader range of deep learning applications.

III Methods

The model in this paper will be divided into two aspects: training and testing, since ultimately the task to be solved in this paper is not to classify the given palm vein images, but to determine whether the given pair of palm vein images belong to the same person. If reconstructing the dataset into pairs of palm vein images and labels of whether they belong to the same person, will bring a large workload and is not conducive to transfer learning using existing convolutional networks. Therefore, after extracting the optical features, we use VGG-16 for feature extraction, send the data to the linear classification layer during training, and use the output given by this layer to calculate the loss in combination with the training labels for training. After the training, the test data is sent into the VGG-16 feature extraction layer to extract features, and then the extracted features are spread into a feature vector, and the similarity of the two image feature vectors is calculated using the similarity function. And a certain threshold value is used to decide whether these two images belong to one person or not. Figure5 gives the general structure of the model in this paper, and each link will be introduced in this paper.

Refer to caption
Figure 1: Model Structure

III-A VGG Transfer Learning

III-A1 Introduction of VGG structure

K Simonyan proposed the VGG deep learning network, A Zisserman in 2014 [18], using an architecture with very small (3x3) convolutional filters for a comprehensive evaluation of networks with increasing depth, and introducing a 1*1 convolutional kernel in the convolutional structure of VGG, without affecting the input-output dimensionality, and introducing nonlinear transformations to increase the expressiveness of the network and reduce the computational effort. They demonstrate that increasing the depth to 16-19 weight layers can significantly improve the existing technical configuration.

VGG-16 has 16 layers, divided into 13 convolutional layers and 3 fully connected layers. The first time, after two convolutions of 64 convolutional kernels, one pooling is used, the second time, after two convolutions of 128 convolutional kernels, another pooling is used, and two more repetitions of three 512 convolutional kernels are convolved and then pooled. Finally, three full connected layers are used to obtain the results. The schematic structure is shown below[51].

Refer to caption
Figure 2: VGG-16

At the end of the network, VGG uses a set of fully connected nodes for the output, where the final output size is the number of categories of images in the dataset. In the middle of each fully connected layer, we add a dropout layer to randomly deactivate a certain percentage of neurons, which avoids the occurrence of model overfitting [42]. Finally, the obtained results are fed into the SoftmaxSoftmax function (1) to find the probability distribution, and finally, the loss is calculated and the error is back-propagated by the cross-entropy loss function (2).

Softmax(zi)=ezic=1Cezc\operatorname{Softmax}\left(z_{i}\right)=\frac{e^{z_{i}}}{\sum_{c=1}^{C}e^{z_{c}}} (1)
L=1NiLi=1Nic=1Myiclog(pic)L=\frac{1}{N}\sum_{i}L_{i}=-\frac{1}{N}\sum_{i}\sum_{c=1}^{M}y_{ic}\log\left(p_{ic}\right) (2)

The softmax function can effectively map the output to a probability distribution space and is particularly effective for solving classification problems.

III-A2 Transfer Learning

As more and more machine learning application scenarios emerge and existing supervised learning that performs better requires large amounts of labeled data, so transfer learning is receiving more and more attention. The goal of transfer learning is to apply the knowledge or patterns learned on a domain or task to a different but related domain or problem [22].

For this paper’s palm vein recognition task, we use a VGG-16 network pre-trained on the ImageNet[21] dataset for transfer learning. imageNet is pre-trained on more than one million images containing more than 20,000 categories. The data volume of the palm vein dataset involved in this paper is small, because there is still some difference between object recognition and palm vein recognition. Considering the above, we freeze the first 30 convolutional layers in VGG-16 with gradient, i.e., they do not generate gradient information during the training process, thus the parameters will not be updated. Instead, the remaining 34 convolutional layers are trained with fully connected layers, which greatly shortens the training time and increases the convergence rate while ensuring prediction accuracy.

III-A3 Attention mechanism

When a human sees a set of information, human neurons will automatically scan the global image based on experience and the current target, and obtain the target area that needs to be focused on, that is, the focus of attention. More attention is then devoted to this region to obtain more details about the target and to suppress other useless information. The following figure is a schematic diagram of the application of the attention mechanism in image processing. For such a journal screenshot, attention is focused on the image and the caption, respectively, under our life experience. [23].

Refer to caption
Figure 3: Attention

A brief description of the attention layer is as follows. First, two hidden states (Encoder and Decoder) are created, which are equivalent to the input and output of the structure. Subsequently, a dot product is performed with each hidden state in Decoder and Encoder, and the result is noted as ScoreScore. Then all the scores are sent to the softmax layer (1), so that the higher the original score of the hidden state, the higher its corresponding probability, thus suppressing that invalid or noisy information.

Next, the hidden state of each Encoder is multiplied by the ScoreScore after softmax, and then all aligned vectors are accumulated and the context vectors are sent to the Decoder to obtain the decoded output. the internal structure of the Attention layer is as follows.

Refer to caption
Figure 4: Attention layer structure

For the palm vein recognition task, we want the neural network to focus more on the features and patterns of the hand, and give a relatively low weight to secondary factors such as shadows and illumination. Therefore, we use the spatial Attention mechanism before image input and the channel Attention mechanism for image feature extraction [24]. Another advantage of the Attention layer is that it is a plug-and-play adaptive module, which does not change the dimensionality of the image and does not change the subject model. It can also be applied to the training of transfer models before the input or after the output of the subject model.

III-B Training and matching algorithms

III-B1 Framework and optimization

In this paper, we build and train the model based on the Pytorch framework [26], and optimize the model using the Adam algorithm [25]. Adam algorithm is different from the traditional stochastic gradient descent which maintains a single learning rate to update all the weights, and the learning rate does not change during the training process. In contrast, the Adam algorithm designs independent adaptive learning rates for different parameters by computing the gradient’s first-order moment estimates and second-order moment estimates. This allows the loss to drop faster in the early learning period without jaggedness in the late learning period due to excessive learning rate.

Meanwhile, we use the (4)L2L_{2} penalty term for the linear output layer, which enables the network to output a relatively sparse feature vector, which is more helpful for the subsequent similarity calculation. The L2L_{2} penalty term can limit the size of the secondary elements in the weights, and in terms of optimization difficulty L2L_{2} penalty is more convex compared to L0L_{0} and L1L_{1} penalties [32], so the optimization is less difficult and the results obtained are better. Also, the inclusion of the regular term helps to avoid overfitting because it compresses part of the parameters to a number close to 0, making the model simpler.

penalty=w2penalty=\left\|w\right\|_{2} (3)

Where w2\left\|w\right\|_{2} is the L2 penalty term, which is easier to compute than the L1 parametrization, and the derivative values are defined at each point. Including the L2 parametrization is achieved simultaneously to facilitate optimisation and obtain a sparse solution. The expression of the loss function after adding the canonical term is as follows.

loss=Crossentropy(X,y)+λw2loss=Crossentropy(X,y)+\lambda\left\|w\right\|_{2} (4)

III-B2 Matching algorithm

After feature extraction by the above deep learning network, for the input iith palm vein image, we can get the feature vector ViV_{i}. Subsequently, to determine whether the iith and jjth images are from the same person, we use the cosine similarity function for matching comparison, whose expression is shown in (5).

s=cos(θ)=ViVjViVj=k=1nVik×Vjkk=1n(Vik)2×k=1n(Vjk)2s=\cos(\theta)=\frac{V_{i}\cdot V_{j}}{\|V_{i}\|\|V_{j}\|}=\frac{\sum_{k=1}^{n}V_{ik}\times V_{jk}}{\sqrt{\sum_{k=1}^{n}\left(V_{ik}\right)^{2}}\times\sqrt{\sum_{k=1}^{n}\left(V_{jk}\right)^{2}}} (5)

This function accepts two feature vectors extracted from the neural network and obtains the similarity by employing a generalised cosine operation. When s>αs>\alpha, we decide that the two palm vein pictures originate from the same person, and reject them in the opposite case. Obviously, in the most ideal case, for any ViV_{i} and VjV_{j} coming from the same person, there should be:

s=cos(Vi,Vj)=1s=\cos(V_{i},V_{j})=1 (6)

And for any ViV_{i} and VjV_{j} from different persons, there should be:

s=cos(Vi,Vj)=0s=\cos(V_{i},V_{j})=0 (7)

But in practical scenarios, we can relax this judgment condition. We usually use the threshold α\alpha to determine whether it is the same person. It follows that as long as the difference between the similarity sis_{i} for any pair of palm veins from the same person and the similarity sjs_{j} for any pair of palm veins from different people is large enough, then we can always find a suitable α\alpha that allows us to distinguish whether a given input is the same person or not. That is, the difference between the similarity values computed by the similarity function from the same person and different persons should be as large as possible, i.e., equation (8) should be as large as possible. We call this equation matching goodness of fit, which is counted as MGMG.

MG=Average(cos(Vis,Vjs))Average(cos(Vid,Vjd))MG=Average(\cos(V_{is},V_{js}))-Average(\cos(V_{id},V_{jd})) (8)

This facilitates us to pick a better threshold α\alpha more easily, making the rate of wrong rejections and wrong acceptance decrease.

III-B3 Multi-task Loss Function

Our loss function is shown in (4) for the traditional classification task. That is, we solve the problem posed by (8) by solving the (4)-style. However, there is no research to prove that there is a relaxation optimization relationship between the two, and it may not be effective if the classification model is more indirect for the matching task. Therefore, we would like to add (8) to the loss function for optimization. The reconstructed loss function is shown as follows. It can be seen as a linear combination of a classification task and a matching task.

loss=θCrossentropy(X,y)+(1θ)(1MG)+λw2loss=\theta*Crossentropy(X,y)+(1-\theta)(1-MG)+\lambda\left\|w\right\|_{2} (9)

Although we want the value of equation (8) to be the largest for all data (i.e., 1MG1-MG is the smallest), it is very detrimental to training because it takes too much time to compute MGMG. However, since we divide the training batches randomly, each training batch 1MG1-MG minimum is equivalent to the overall 1MG1-MG minimum.

The parameter θ\theta controls the weight between two tasks, θ=1\theta=1 for a pure classification task and θ=0\theta=0 for a pure matching task. It is worth noting that the premise of computing the batch MGMG is that there must be both data of the same category and data of different categories in each batch. To achieve this, we may need to increase the data in each batch, which inevitably leads to a decrease in training accuracy. We will elaborate on the specific parameter selection in Chapter 3.

IV Experiment and Result

The following is the training environment for this article.

  • \bullet

    Python 3.9.12

  • \bullet

    Pytorch 1.11.0

  • \bullet

    CUDA 11.6

  • \bullet

    Intel(R) Xeon(R) Gold 5320 CPU @ 2.20GHz

  • \bullet

    Tesla V100*1

  • \bullet

    32GB RAM

To simulate a realistic application, usually without GPU acceleration, here is the test environment for this article.

  • \bullet

    Macbook pro

  • \bullet

    Apple silicon M1 Pro

  • \bullet

    16GB RAM

IV-A Datasets

There are currently five commonly used 2D palmprint databases in the academic community, including the Hong Kong Polytechnic University’s Palmprint II database (PolyU II), the Hong Kong Polytechnic University’s Blue Ribbon Multispectral database (PolyU M_B)[46], the Hefei University of Technology’s Palmprint database (HFUT)[47], the Hefei University of Technology’s Cross-Sensor Palmprint database (HFUT CS)[48]. The Tongji University Palmprint database (TJU-P)[49]. The dataset used in this paper is PolyU M_B and TJU-P [34][43]. The datasets from TJU-P have been centered so that all the palm vein areas are recognizable on the screen, while the PolyU M_B datasets are the entire palm, which may make prediction more difficult, however, at the same time, more common in real life.

Refer to caption
Figure 5: A sample from the two datasets

These databases generally contain two parts, derived from the same samples, a period of time before and after palm print or palm vein sampling. When using deep learning methods for palm print and palm vein recognition, it is easy to improve recognition accuracy if the training set contains samples from both parts, and this experimental mode is referred to as the mixed data mode. If the training set is only from the first part and the test samples are from the second part, the training pattern is referred to as the separated data pattern[50]. In this experiment, the separated data model was used for this task in order to obtain a more accurate and realistic recognition rate.

IV-B Data pre-processing

IV-B1 Image size processing

To fit the input size of the convolutional network, we first cropped the image to 224*224 centered on the palm to be the ROI(region of interest). The images were sourced from the same person by recording 6 consecutive images and 6 images 10 days later. That is, there are 12 pictures from the same hand each. Therefore, we use the former as the training set, counting as the set TRTR for the classification task. And every 9 photos are taken out as the validation set, which is counted as the set VV. For the matching process, in order to simulate the entry-comparison process in real application scenarios, we extract the last 10 sets of photos (i.e., 10 different palms) from the dataset as the test set for zero-experience learning matching to test the matching effect, and match the photos in the two datasets one by one to form a 60*60 dataset, which is counted as The set TEmTE_{m}.

IV-B2 Optical pre-processing

Image enhancement is mainly used to highlight certain information or to meet specific needs. Techniques for image enhancement generally fall into two main categories: null domain image enhancement and frequency domain image enhancement algorithms, in addition to image enhancement algorithms that combine fuzzy theory and genetics.

Space domain image enhancement is the direct processing of image pixels and is essentially based on a grey-scale mapping transformation. Commonly used methods include histogram equalization, spatial filtering, and grey scale transformations.

The essence of frequency domain image enhancement is convolution. It mainly uses the Fourier transform or wavelet transform to transform the image to the frequency domain, then frequency domain enhancement, and finally inverse transformation back to the null domain to obtain the desired image. Common methods include low-pass filtering, high-pass filtering, etc.

Enhancement of palm print images can eliminate noise in the original image, balance illumination effects and enhance contrast. This paper selects four image enhancement methods: histogram equalization, Laplace operator, Constrained Contrast Adaptive Histogram Equalisation algorithm (CLAHE), and log transform. Two samples were randomly selected on two datasets respectively and processed as shown in the following figures .

Refer to caption
Figure 6: Impact of four optical pretreatments on the TJU-P dataset
Refer to caption
Figure 7: Impact of four optical pretreatments on the PoLy-U M_N dataset

Among them, the log transform was the least effective on the TJU-P dataset, while the laplacian, clahe, and hist effectively removed the noise from the original images. On the PolyU M_N dataset, the laplacian operator processed results do not differ much from the original image, while the remaining methods each highlight distinct features after enhancement.

IV-C Judging criteria

For a single pair of samples, we can use the deviation from the ideal state equation (6, 7) to measure the deviation. Since the matching result is essentially a dichotomous variable of ”acceptance” and ”rejection” for a given multi-treatment matching sample, we use the AUC index to judge the matching merit. The following formula calculates it.

AUC=predpos>prednegpositiveNumnegativeNumAUC=\frac{\sum{pred}_{{pos}}>{pred}_{{neg}}}{{positiveNum*negativeNum}} (10)

It is noted that positive and negative cases may be unbalanced in this problem. In contrast, the calculation of AUC considers the classification ability of positive and negative cases and can reasonably evaluate the classifier despite the unbalanced sample. Therefore, AUC is not sensitive to whether the sample categories are balanced or not, and is suitable as a scoring criterion for this work.

IV-D Ablation Study

In this section, we test the effects of two techniques: optical enhancement and attention mechanism, on the model’s performance before and after their introduction. Comparative experiments on multi-task loss will be shown later.

IV-D1 Optical Enhancement Comparison

We compared several optical enhancements mentioned in II. B. 2) and tested them in both datasets with the following results.

On the TJU-P dataset, the enhanced palm prints of the images by three different ways were compared with the results of the original images after 15 training sessions. And the batch size is 64, the learning rate is 5×1055\times 10^{-5} and the optimiser is Adam. The results for loss and matching accuracy are shown in Figure 9 and Figure 9.

Refer to caption
Figure 8: Loss of different enhancement
Refer to caption
Figure 9: Accuracy of different enhancement

From the experiments, it can be seen that after image enhancement by histogram equalization, loss converges the fastest and has the highest recognition rate improvement; CLAHE has the second highest effect; Laplace operator processing has the least significant effect, and the accuracy rate is even lower than the recognition rate of the unprocessed image.

On the PolyU M_N palm vein dataset, only these two image enhancement methods were used for comparison as hist and CLAHE were shown to achieve better results through the above experiments. The parameters were chosen: the number of iterations was 15, the batch size was 8, and the learning rate and optimiser were the same as before. The comparison results are shown below in Figure 11 and Figure 11.

Refer to caption
Figure 10: loss
Refer to caption
Figure 11: accuracy

As can be seen from the figure: on the PolyU M_N palm vein dataset, the enhancement effect of image enhancement using hist and clahe is almost the same, and hist is slightly better in terms of convergence speed and accuracy.

IV-D2 Attention mechanisms

We then tested the effect of introducing attentional mechanisms on the experimental results. Using the Poly-U M_N dataset as an example, we tested the matching accuracy with and without the inclusion of the attention mechanism. The results are presented in Figure 12.

Refer to caption
Figure 12: Effect of attention on accuracy

The results show that the attention mechanism’s introduction largely improves convergence speed. It also improves the accuracy of the model at convergence.

The attention mechanism can be used to weight different spatial regions or channels, as can be seen from the weighting of channels by the attention mechanism in the Figure 13, where the weight of some channels is even set to 0. This also serves the purpose of feature selection. Since we found that the attention mechanism performed well in the model, we will use attention as the underlying structure in all subsequent experiments.

Refer to caption
Figure 13: Effect of attention on accuracy

IV-E Results of Classification-matching Methods

IV-E1 Classification Results

Batch size = 32 is set, 20 rounds are trained, and a regularization strength of λ=0.001\lambda=0.001 is used. The resulting classification accuracies and loss functions for the training set, validation set and test set are shown in Figure 15 and Figure15, and the training was stopped at the 7th round due to the fast convergence rate.

Refer to caption
Figure 14: Train loss
Refer to caption
Figure 15: Train accuracy

The training took a total of 173 seconds for 7 rounds, and it can be found from the figure that the model converged well. The accuracy is 100% on the training set and 99.85% on the validation set. We also tried different combinations of parameters, and the results are shown in the table 1.

Batch Size Epoch Regularization strength Loss Train Accuration Test Accuration Early stop Training Time
32 10 0.001 0.000385 100% 99.85% 7 15m32s
64 10 0.001 0.000461 100% 98.73% No 12m14s
32 5 0.001 0.000612 99.63% 98.77% No 7m54s
64 5 0.001 0.001324 96% 95.87% No 5m59s
32 10 0.01 0.00836 95.42% 94.68% 8 15m17s
64 10 0.01 0.00972 95.11% 94.32% No 13m05s
32 5 0.01 0.0132 94.91% 89.23% No 8m16s
64 5 0.01 0.0187 91.61% 85.09% No 6m28s
TABLE I: Experimental results of the original loss function

The results show that with the regularization strength of λ=0.001\lambda=0.001, the final loss and accuracy are not sensitive to the number of batches and training sessions. However, for λ=0.01\lambda=0.01 the model appears to be underfitted.

At the same time, we found that too large batch size will greatly affect the training effect. Although a large batch size helps to speed up training, with the support of a powerful GPU, the time consumption here is negligible.

IV-E2 Matching Results

Next, the matching effect is tested on the dataset TEmTE_{m}. Also, to get a better matching effect, we test thresholds of 0.6, 0.65, 0.7, 0.75, and 0.8, calculate their AUCs and compare them. Figure 16 shows the comparison graph of AUC, and it can be found that the best result is obtained when the threshold is 0.6. Figure 18 shows the prediction when the threshold is 0.6, where yellow is the correct prediction and blue is the wrong prediction.

Refer to caption
Figure 16: Comparison of AUC at different thresholds

The overall correct rate was 95.08%. Among them, 279 pairs were correctly accepted, 121 pairs were incorrectly rejected, 56 were incorrectly accepted, and 3144 were correctly rejected. Although the overall accuracy rate is very high, the acceptance rate for positive samples is very low, even less than 50%. Considering that the classification accuracy has reached 100%, we need to modify the loss function in the training to further improve the matching accuracy.

Meanwhile, we used the (8) equation to calculate the goodness-of-fit under the matching task, i.e., the difference between the average similarity from the same sample and the average similarity from different samples were calculated separately. The experimental results show that the average similarity of samples from the same palm is 0.7700, and the average similarity of samples from different palms is 0.2531, with a difference of 0.5169, which is the MGMG. We find that the correct acceptance rate of this method is low and far from the standard for practical use, so we will try our improved objective function next.

IV-F Result based on multi-tasks loss function

Set Batch size = 128, keep the rest parameters the same as the optimal parameters above, change the value of θ\theta for several experiments, and take one of the training accuracy and loss variation graphs to show as follows.

Refer to caption
Figure 17: multi-tasks loss function loss
Refer to caption
Figure 18: multi-tasks loss function accuracy

We found that the model converged slower due to the improvement of the loss function. The model stopped training at 12 rounds and took 1247 seconds. The loss function’s training time is higher than the normal classification task. We tried different combinations of parameters and the obtained experimental results are shown in the table 2.

Theta Alpha Correctly accepted Wrong accepted Wrong rejected Correctly rejected Correct Rate AUC MG
0.5 0.7 310 132 36 3122 95.33% 0.9442 0.7942
0.3 0.7 345 79 15 3161 97.39% 0.9669 0.8293
0.1 0.7 324 79 36 3161 96.81% 0.9378 0.8121
TABLE II: Experimental results of multi-tasks loss function

Where α\alpha are used to select the optimal value by the above method and change the value of θ\theta, the results show that the highest matching accuracy is achieved when θ=0.3\theta=0.3, when the loss function is as follows.

loss=0.3Crossentropy(X,y)+0.7(1MG)+λw2loss=0.3*Crossentropy(X,y)+0.7*(1-MG)+\lambda\left\|w\right\|_{2} (11)

The following Figure 20 and Figure 20 shows the experimental results of the original loss function and the multi task loss function on the matching task. The yellow data points represent correctness and the blue data points represent errors.

Refer to caption
Figure 19: Original loss function prediction results
Refer to caption
Figure 20: multi-tasks loss function prediction results

It can be seen that the multi-task loss function proposed in this paper has a higher prediction accuracy. We also test the goodness of matching, the difference between the average similarity of samples from different palms and the same palms. It is 0.8293 which is far higher than the result in section A.

IV-G The case where the training and test sets are from different datasets

The palm vein images used in the above model training and matching are from the same dataset. In other words, the locations and parameters of the photos taken are approximately the same. Testing these photos alone does not measure the true robustness and transferability of the model. Therefore, we next use photos taken in different ways for matching tests. To improve the model’s accuracy, we use palm vein images taken in two completely different environments for training, and use the dataset from the third case for matching tests. The appearance of the training image and the test image are shown below.

Refer to caption
Figure 21: Different training and test sets

The palm veins are more informative in the top left image, while the palm surface is more informative in the bottom left image. And in this case, the determination of the matching threshold is key. Since the test and training sets are from very different datasets, the positive and negative examples cannot be simply partitioned by a number like 0.6 or 0.7. Therefore, we must try to determine the matching threshold α\alpha. From the above discussion, we can find that based on the Multi-tasks loss function, we believe that the final set of obtained feature vectors should satisfy the MGMG(8) minimum. In this case, we believe that the similarity of positive and negative examples should each be well separated, and although they will partially cross, we can always find a better threshold based on such an assumption. Therefore, we first perform K-mean[45] clustering on the similarity of all samples and get two clustering centers pp and nn, and we take α=p+n2\alpha=\frac{p+n}{2} as the threshold for matching.

In the case of the experiments in this paper, we finally choose α=0.992\alpha=0.992. Predictions were performed for a test set with 46 pairs of positive cases and 134 pairs of negative cases. Out of 180 pairs of predictions, only 2 pairs of palms that should have belonged to the same person were incorrectly rejected. The overall accuracy rate was 98.89%. The following figure shows the confusion matrix of the predicted results.

Refer to caption
Figure 22: Confusion matrix of test results

We find that the model still shows very high accuracy on different datasets, so we consider the training idea of the palm vein matching model we have provided is relocatable. At the same time, we compared the accuracy of the validation set under different methods and the results obtained are shown in the Table 3. The results show that the VGG network with attention mechanism and multi-task loss improves 11.36% in accuracy over the original VGG network, and 5.11% over the attention mechanism alone.

Model Correctly accepted Wrong accepted Wrong rejected Correctly rejected Correct Rate AUC MG
VGG 115 7 15 39 87.50% 0.8662 0.7301
Attention-VGG 124 5 6 124 93.75% 0.9226 0.8271
multi-task-Attention-VGG 130 2 0 44 98.86% 0.9783 0.8724
TABLE III: Accuracy comparison on the validation set

IV-H Simulation of reality: noise, displacement and low resolution

The palm print dataset in the above experiments was collected in a very regular pattern and they used a limiter to fix the position of the palm. However, in real-life application, the position of the hand, the environment, and the resolution of the acquisition device may all change. Therefore, we rotate, blur and randomly noise the palmprint images of the validation set. And experiments are conducted in the same way as above to explore the robustness of the model. Figure 23 shows a sample of images with the four processing methods

Refer to caption
Figure 23: Simulation of realistic image processing

We did this for each image in the validation set and compared the AUC of the different treatments. The results are shown in Table 4.

Treatment Correct Rate AUC MG
Original 98.86% 97.83% 0.8724
Blur 57.16% 53.44% 0.4139
Noise 93.72% 92.73% 0.8322
Rotation 92.14% 91.09% 0.8147
TABLE IV: Results of experiments in simulating realistic scenarios

We can find that matching accuracy decreases in all three unique treatments. The Noise and Rotate drop is insignificant and still has an accuracy rate of over 90%. On the other hand, the model has almost no recognition ability for the blur. Our analysis is that this is due to blurring resulting in the complete loss of otherwise obscure palm print features, which makes samples from the same individual look the same as samples from different individuals. This is illustrated by MG, which is only 0.4139. This indicates that the similarity of the two types of samples has a considerable overlap, which means that a threshold cannot simply split it.

The above experimental results show that this method is robust to both common noise and rotation. The loss of features due to blurring means that there is a requirement for the device’s resolution in this recognition mode.

IV-I Time consumption

We also test the time complexity of the matching process. Considering that most of the application scenarios are not supported by GPU arithmetic, in order to simulate real usage scenarios, the test will be based on the Apple M1 Pro platform using CPU for inference. In the test on 3600 pairs of samples, the average test time per pair of samples was 0.1316 seconds, and the time consumption curve for each pair of samples is shown in Figure 24.

Refer to caption
Figure 24: Prediction time consumption

The results show that the algorithm is short time consuming and stable, which has the possibility to be used in real life.

V Conclusion, Discussion and Future Work

V-A Innovation

In this paper, we combine optical feature extraction with depth feature extraction from the traditional method of optical recognition of palm veins. Through experiments, it is found that this approach can better mine image texture features and provide features that can be multi-dimensional for matching analysis. At the same time, the use of VGG-16 transfer learning improves the learning efficiency and makes the model training overhead much lower; the practice of freezing part of the pre-training parameters makes use of the original image recognition ability of the network on the one hand and can be better adapted to different tasks on the other. At the same time, we introduce the Attention mechanism, which enables the model to distinguish between primary and secondary features after feature extraction, improving the model robustness and further increasing the model accuracy. Finally, we use the trained model for the palm vein matching task and get an accuracy rate of more than 98%.

At the same time, we propose an index to evaluate the training results’ average Similarity, which can represent the distance of similarity between different kinds of palmprint images, to describe the accuracy and difficulty of distinguishing them. In the future study, we can use this metric to evaluate and compare traditional matching solutions, such as using PCA to extract features by dimensionality reduction and using the SVM support vector machine to match [41]

Finally, to satisfy the migration of the model to other datasets. We determine the adaptive matching threshold based on a clustering algorithm, and the experimental results prove that the method is simple and feasible with high accuracy.

Refer to caption
Figure 25: Innovations

V-B Shortcomings and future work

However, in this paper, no special treatment is done for the dataset. A random method is still used for the division of batches. In future work, we can reconstruct the dataset so that the number of data from the same sample and different samples in each batch is approximately the same. This would be more representative and reliable for the MGMG calculation, thus further improving the model performance. Meanwhile, this paper’s VGG network used for migration learning is not the best image recognition network available. There are well-performing networks such as Resnet. The latest results are given by NFNet[44] proposed by DeepMind.

A similar improvement to this idea was proposed by Google at CVPR 2015 [37], which reconstructs the face classification dataset and proposes Triplets Loss for training. This network is called FaceNet, and its ultimate goal is to embed face images into a 128-dimensional Euclidean space and get a better matching result by minimizing the Euclidean distance between individuals and maximizing the Euclidean distance between different individuals.

Furthermore, due to the lack of infrared devices, this paper has not yet conducted demo production and application tests on the matching of the model, i.e., its performance in continuous and real-time recognition has not been tested. These aspects need to be further explored and studied.

References

  • [1] D. Zhang, Automated Biometrics: Technologies and Systems, New York, NY, USA:Springer, 2013.
  • [2] A. K. Jain, A. Ross and S. Prabhakar, ”An introduction to biometric recognition”, IEEE Trans. Circuits Syst. Video Technol., vol. 14, no. 1, pp. 4-20, Jan. 2004.
  • [3] F. Chen, X. Huang and J. Zhou, ”Hierarchical minutiae matching for fingerprint and palmprint identification”, IEEE Trans. Image Process., vol. 22, no. 12, pp. 4964-4971, Dec. 2013.
  • [4] Y. Ding, D. Zhuang and K. Wang, ”A study of hand vein recognition method”, Proc. IEEE Int. Conf. Mechatronics Autom., pp. 2106-2110, 2005.
  • [5] X. Wang and X. Tang, ”A unified framework for subspace face recognition”, IEEE Trans. Pattern Anal. Mach. Intell., vol. 26, no. 9, pp. 1222-1228, Sep. 2004.
  • [6] M. I. Ahmad, W. L. Woo and S. Dlay, ”Non-stationary feature fusion of face and palmprint multimodal biometrics”, Neurocomputing, vol. 177, pp. 49-61, Feb. 2016.
  • [7] Galbally, J., et al. ”Image Quality Assessment for Fake Biometric Detection: Application to Iris, Fingerprint, and Face Recognition.” Ieee Transactions on Image Processing 23(2): 710-724, 2014.
  • [8] Lu, G. M., et al. ”Palmprint recognition using eigenpalms features.” Pattern Recognition Letters 24(9-10): 1463-1467, 2003.
  • [9] D.-S. Huang, W. Jia and D. Zhang, ”Palmprint verification based on principal lines”, Pattern Recognit., vol. 41, no. 4, pp. 1316-1328, Apr. 2008.
  • [10] X. Wu, D. Zhang and K. Wang, ”Palm line extraction and matching for personal authentication”, IEEE Trans. Syst. Man Cybern. A Syst. Humans, vol. 36, no. 5, pp. 978-987, Sep. 2006.
  • [11] Jain, A. K. and J. J. Feng. ”Latent Palmprint Matching.” Ieee Transactions on Pattern Analysis and Machine Intelligence 31(6): 1032-1047, 2003.
  • [12] Zhao, S. P. and B. Zhang. ”Joint Constrained Least-Square Regression With Deep Convolutional Feature for Palmprint Recognition.” Ieee Transactions on Systems Man Cybernetics-Systems 52(1): 511-522, 2022.
  • [13] Trabelsi, S., et al. ”Efficient palmprint biometric identification systems using deep learning and feature selection methods.” Neural Computing & Applications, 2022.
  • [14] Shin, H. C., et al. ”Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning.” Ieee Transactions on Medical Imaging 35(5): 1285-1298, 2016.
  • [15] Kiselev, V. Y., et al. ”Challenges in unsupervised clustering of single-cell RNA-seq data.” Nature Reviews Genetics 20(5): 273-282, 2019.
  • [16] Yosinski, J., et al. ”How transferable are features in deep neural networks?” Advances in neural information processing systems 27, 2014.
  • [17] He, K., et al. Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition, 2016.
  • [18] Simonyan, K. and A. Zisserman. ”Very deep convolutional networks for large-scale image recognition.” arXiv preprint arXiv:1409.1556, 2014.
  • [19] Pizer, S. M., et al. ”Adaptive histogram equalization and its variations.” Computer vision, graphics, and image processing 39(3): 355-368, 1987.
  • [20] LeCun, Y., et al. ”Gradient-based learning applied to document recognition.” Proceedings of the IEEE 86(11): 2278-2324, 1998.
  • [21] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li and L. Fei-Fei, ImageNet: A Large-Scale Hierarchical Image Database. IEEE Computer Vision and Pattern Recognition (CVPR), 2009.
  • [22] Pan, S. J., et al. ”Domain adaptation via transfer component analysis.” IEEE transactions on neural networks 22(2): 199-210, 2010.
  • [23] Vaswani, A., et al. ”Attention is all you need.” Advances in neural information processing systems 30, 2017.
  • [24] Choi, M., et al. Channel attention is all you need for video frame interpolation. Proceedings of the AAAI Conference on Artificial Intelligence, 2020.
  • [25] Kingma, D. P. and J. Ba. ”Adam: A method for stochastic optimization.” arXiv preprint arXiv:1412.6980, 2014.
  • [26] Paszke, Adam and Gross, Sam and Chintala, Soumith and Chanan, Gregory and Yang, Edward and DeVito, Zachary and Lin, Zeming and Desmaison, Alban and Antiga, Luca and Lerer, Adam. ”Automatic differentiation in PyTorch.” NIPS-W, 2017.
  • [27] Vinoth, R. and C. Venkatesh. Segmentation and Detection of Tumor in MRI images Using CNN and SVM Classification. 2018 Conference on Emerging Devices and Smart Systems (ICEDSS), IEEE, 2018.
  • [28] Huang, X., et al. ”Identifying disaster related social media for rapid response: a visual-textual fused CNN architecture.” International Journal of Digital Earth, 2019
  • [29] N. Duta, A.K. Jain, K.V. Mardia Matching of palmprints Pattern Recognition Letters, 23 (4), pp. 477-485, 2002
  • [30] W. Shu, D. Zhang Automated personal identification by palmprint Optical Engineering, 38 (8), pp. 2359-2362, 1998.
  • [31] Kong, A., et al. ”A survey of palmprint recognition.” Pattern Recognition 42(7): 1408-1418, 2009.
  • [32] Candès, E. J. and B. Recht. ”Exact matrix completion via convex optimization.” Foundations of Computational mathematics 9(6): 717-772, 2009.
  • [33] Jia, Wei & Xia, Wei & Zhao, Yang & Min, Hai & Chen, Yan-Xiang. 2D and 3D Palmprint and Palm Vein Recognition Based on Neural Architecture Search. International Journal of Automation and Computing. 18. 10.1007/s11633-021-1292-1, 2021.
  • [34] A. Genovese, V. Piuri, K. N. Plataniotis, F. Scotti. PalmNet: Gabor-PCA convolutional networks for touchless palmprint recognition. IEEE Transactions on Information Forensics and Security, vol.14, no.12, pp.3160–3174, 2019.
  • [35] Han, W.-Y. and J.-C. Lee. ”Palm vein recognition using adaptive Gabor filter.” Expert Systems with Applications 39(18): 13225-13234, 2012.
  • [36] Weldon, T. P., et al. ”Efficient Gabor filter design for texture segmentation.” Pattern Recognition 29(12): 2005-2015, 1996.
  • [37] Schroff, F., et al. Facenet: A unified embedding for face recognition and clustering. Proceedings of the IEEE conference on computer vision and pattern recognition, 2015.
  • [38] Fanjiang, Y. Y., et al. ”Palm Vein Recognition Based on Convolutional Neural Network.” Informatica 32(4): 687-708, 2021.
  • [39] Wilson, C. Vein pattern recognition: a privacy-enhancing biometric, CRC press, 2010.
  • [40] Wu, K. S., et al. ”A secure palm vein recognition system.” Journal of Systems and Software 86(11): 2870-2876, 2013.
  • [41] Faruqe, M. O. and M. A. M. Hasan. Face recognition using PCA and SVM. 2009 3rd international conference on anti-counterfeiting, security, and identification in communication, IEEE, 2009.
  • [42] Baldi, P. and P. J. Sadowski. ”Understanding dropout.” Advances in neural information processing systems 26, 2013.
  • [43] PolyU Palmprint database, http://www.comp.polyu.edu.hk/ biometrics/
  • [44] Brock, A., et al. High-performance large-scale image recognition without normalization. International Conference on Machine Learning, PMLR, 2021.
  • [45] MacQueen, J. Classification and analysis of multivariate observations. 5th Berkeley Symp. Math. Statist. Probability, 1967.
  • [46] V. Kanhangad, A. Kumar, D. Zhang, “A unified framework for contactless hand identification,” IEEE Trans. Information Forensics & Security, vol. 20, no. 5, pp. 1415-1424, May 2011.
  • [47] W. Jia, B. Zhang, J. T. Lu, Y. H. Zhu, Y. Zhao, W. M. Zuo, H. B. Ling. Palmprint recognition based on complete direction representation. IEEE Transactions on Image Processing, vol.26, no.9, pp.4483–4498, 2017.
  • [48] W. Jia, R. X. Hu, J. Gui, Y. Zhao, X. M. Ren. Palmprint recognition across different devices. Sensors, vol.12, no.6, pp.7938–7964, 2012.
  • [49] L. Zhang, L. D. Li, A. Q. Yang, Y. Shen, M. Yang. Towards contactless palmprint recognition: A novel device, a new benchmark, and a collaborative representation based identification approach. Pattern Recognition, vol.69, pp.199–212, 2017.
  • [50] Jia, W., Gao, J., Xia, W. et al. A Performance Evaluation of Classic Convolutional Neural Networks for 2D and 3D Palmprint and Palm Vein Recognition. Int. J. Autom. Comput. 18, 18–44, 2021.
  • [51] Sugata, T. L. I., and C. K. Yang. ”Leaf App: Leaf recognition with deep convolutional neural networks.” IOP Conference Series: Materials Science and Engineering. Vol. 273. No. 1. IOP Publishing, 2017.
\EOD