IFViT: Interpretable Fixed-Length Representation for Fingerprint Matching via Vision Transformer
Abstract
Determining dense feature points on fingerprints used in constructing deep fixed-length representations for accurate matching, particularly at the pixel level, is of significant interest. To explore the interpretability of fingerprint matching, we propose a multi-stage interpretable fingerprint matching network, namely Interpretable Fixed-length Representation for Fingerprint Matching via Vision Transformer (IFViT), which consists of two primary modules. The first module, an interpretable dense registration module, establishes a Vision Transformer (ViT)-based Siamese Network to capture long-range dependencies and the global context in fingerprint pairs. It provides interpretable dense pixel-wise correspondences of feature points for fingerprint alignment and enhances the interpretability in the subsequent matching stage. The second module takes into account both local and global representations of the aligned fingerprint pair to achieve an interpretable fixed-length representation extraction and matching. It employs the ViTs trained in the first module with the additional fully connected layer and retrains them to simultaneously produce the discriminative fixed-length representation and interpretable dense pixel-wise correspondences of feature points. Extensive experimental results on diverse publicly available fingerprint databases demonstrate that the proposed framework not only exhibits superior performance on dense registration and matching but also significantly promotes the interpretability in deep fixed-length representations-based fingerprint matching.
Index Terms:
Interpretable Fingerprint Recognition, Vision Transformers, Fingerprint Registration and Matching, Fixed-Length Fingerprint RepresentationI Introduction

FingerPrint is an immutable and unique biological trait widely used for human authentication in various scenarios including forensics, bank identification and physical access control [1, 2]. As a crucial part of authentication, fingerprint matching aims to compare the input fingerprint patterns with those stored in a database to determine if they belong to the same finger. Minutiae, e.g. ridge endings and bifurcations, are commonly considered reliable features for accomplishing the matching process. However, extracting minutiae may be challenging when fingerprint quality is low due to conditions such as dry or wet [3]. Conversely, deep learning-based approaches are capable of extracting discriminative fixed-length fingerprint representation and have been considered a promising alternative to address the limitations of traditional minutiae-based matching methods [3, 4, 5, 6]. Despite significant progress, the improvement of interpretability in deep learning-based fingerprint matching is still in its infancy.
Machine Learning (ML) methods have gained tremendous success in major fields due to their powerful inferential capabilities [7, 8, 9, 10, 11], among which explainable Artificial Intelligence (XAI) is currently one of the key focuses [12, 13]. XAI aims to enhance the comprehensibility and transparency on the outcomes of artificial intelligence systems to facilitate reliable real-world data-driven applications. Understanding the underlying reasons behind ML decision-making is crucial, particularly for black-box deep learning models. For fingerprint matching, accurately determining feature points at the pixel level used for matching has the potential to significantly improve the interpretability. Although conventional minutiae-based matching approaches can establish pixel-wise correspondences of minutiae between an input fingerprint pair [14], performing pairwise comparisons of fingerprints is computation-intensive and varies with the number of detected minutiae, especially in the encrypted domain [3]. To overcome this challenge, learning fixed-length fingerprint representation has emerged as a promising solution [3]. However, directly revealing pixel-wise correspondences of feature points within the provided fingerprint pair remains challenging.
Deep learning models have recently been explored to build correspondences for a given fingerprint pair. Cao et al. [5] introduced the Autoencoder-based model for minutiae detection and established minutiae correspondences for computing the similarity between latent and reference fingerprints. To mitigate the negative impact of nonlinear skin distortion, fingerprint dense registration techniques [15, 16] are proposed to measure pixel-wise displacement between two fingerprints. However, such models primarily serve as the basis for subsequent tasks, e.g. fingerprint matching or mosaicking [15], and thus cannot directly provide interpretable pixel-wise correspondences of feature points in the matching result based on the model’s output, such as fixed-length representations. Grosz et al. [17] devise AFRNet to extract discriminative fingerprint representations for fingerprint indexing, enabling visualization and refinement using correspondences between local fixed-length fingerprint representations in low-certainty situations. However, the convolutional neural network branch of AFRNet limits the number of obtainable correspondences due to its local focus. Additionally, the correspondences produced by AFRNet are patch-based rather than pixel-wise and are derived through a computationally intensive brute-force algorithm. Given these considerations, it is imperative to develop a reliable approach that can learn fixed-length representations and provide interpretable dense pixel-wise correspondences of feature points simultaneously, especially in low-texture areas of fingerprints.
Existing architectures for extracting fixed-length fingerprint representation are primarily based on Convolutional Neural Networks (CNNs), often accompanied by domain knowledge (e.g. minutiae information) [3, 18, 6] to improve the matching performance. Other than CNNs, the application of Vision Transformer (ViT) has also emerged recently as an effective solution [19, 17], owing to its superior performance in comparison to CNNs in various computer vision tasks [20, 21]. Unlike CNNs focusing on local regions, the ViT utilizes a self-attention mechanism to attend to various parts of an input image, enabling it to capture long-range dependencies and global context [22, 23]. Such characteristics may be particularly useful for extracting features pertaining to long-range texture information, e.g. ridges and valleys of fingerprints.

To the best of our knowledge, there is no deep learning-based fingerprint matching method that could simultaneously produce interpretable dense pixel-wise correspondences of feature points and discriminative fixed-length representations directly for the given fingerprint pair. Instead of primarily aiming to improve recognition performance like previous studies [24, 3, 17], we focus more on improving the interpretability of fingerprint matching whilst maintaining reasonable recognition performance. Inspired by the global attention mechanism of ViT, we propose a multi-stage fingerprint matching network, namely Interpretable Fixed-length Representation for Fingerprint Matching via Vision Transformer (IFViT), to learn interpretable fixed-length representation to capture long-range relationships between different parts of the fingerprint. Fig. 2 presents the entire architecture of IFViT. It is composed of two modules, namely an interpretable fingerprint dense registration module and an interpretable fixed-length representation extraction and matching module. In contrast to the previous fingerprint dense registration approaches implemented by CNN [15, 16], this study is the first to utilize the ViT with the global attention mechanism for producing dense pixel-wise correspondences of feature points. In the challenging examples shown in Fig. 1, a large number of correspondences of feature points could be still produced in the case of low-quality and cross-sensor fingerprint pairs based on the proposed method. Additionally, we also develop a ViT-based Siamese Network as the primary backbone for fingerprint matching, taking into account both local and global representations of fingerprints to learn fixed-length representations and provide interpretable pixel-wise correspondences of feature points in matching results.
We evaluate our proposed model on several public fingerprint datasets against previous deep learning-based fingerprint recognition methods. Experimental results demonstrate its effectiveness in terms of registration and matching performances. Moreover, the proposed method could offer interpretable dense pixel-wise correspondences of feature points in the alignment and matching results, a distinctive feature absent in previous works on learning fingerprint fixed-length representations. We summarise our contributions as follows:
-
•
We introduce a ViT-based interpretable multi-stage deep learning framework for fingerprint matching. It could not only learn the discriminative fixed-length representation but also provide interpretable dense pixel-wise correspondences of feature points on the given fingerprint pair in matching results.
-
•
We propose an interpretable Siamese network-based fingerprint dense registration module via ViT, which can provide interpretable dense pixel-wise correspondences of feature points in the fingerprint pair, even for low-texture areas. These learned correspondences can be further employed for effective fingerprint alignment and improving interpretability in the fingerprint matching stage.
-
•
We propose an interpretable Siamese network-based fixed-length representation extraction and matching module via ViT. It combines local and global representations from fingerprints for semantic shortcomings of each other. The matching performance has been evaluated on diverse public databases and is comparable to state-of-the-art fingerprint recognition methods based on deep learning, when the model is trained on a limited number of fingers.
II Related Work
II-A Fixed-Length Representations of Fingerprints
A number of studies [3, 4, 5, 6, 25] have investigated the feasibility of learning deep fixed-length fingerprint representations in fingerprint recognition. CNN-based approaches are among the first that were widely employed [24, 3, 4, 18]. As the pioneering work, Cao et al. [24] design an Inception V3-based approach for fingerprint indexing and matching using the extracted fixed-length representation [24]. To improve the recognition performance, the domain knowledge e.g. minutiae information is further introduced in deep networks. One of the representative works is DeepPrint [3], which could achieve competitive performance compared to the state-of-art minutiae matchers. The learning of fixed-length representation is also taken into account in contactless and contact fingerprint matching [4]. Different from the aforementioned approaches that are based on classification, similarity metric learning has been also used to learn fixed-length representations. Lin et al. develop a Siamese CNN-based framework for accurately matching contactless and contact-based fingerprint images [18].
More recently, the application of the ViT for learning fixed-length fingerprint representation has also drawn attention [25, 19]. Grosz et al. [25] propose the first use of the ViT in extracting fixed-length fingerprint representation and demonstrate that the independent performance of ViT in fingerprint recognition rivals that of CNN-based models. Tandon et al. [19] propose a convolutional ViT network based on both global representation (i.e. complete fingerprint image) and local representation (i.e. minutiae information) to complement semantic information in fingerprint recognition.
However, most of above-mentioned approaches encounter the limitation of interpretability during the fingerprint matching process. The difficulty lies in the lack of ability of these learned fixed-length representations to determine feature points of the fingerprint pair utilized for matching. They lack the same level of interpretability exhibited by the minutiae-based matching method as the latter can visually provide pixel-wise minutiae correspondences in the matching result.
II-B Interpretability of Fingerprint Matching
Establishing deep learning models with acceptable interpretability has become an important task for numerous studies when dealing with computer vision-related tasks in recent years [12, 26]. As another prevalent biometric technology for identifying individuals, facial recognition has received significant attention, leading to considerable research on interpretable deep learning models for face matching and indexing [27, 28]. In contrast, there is hardly any research on interpretable fingerprint recognition technology. The need for interpretable fingerprint recognition approaches based on deep learning models is equally significant and warrants more consideration.
The minutiae-based fingerprint matching approaches have held a dominant position over the past few decades [29, 30]. This is attributed not only to its reliable performance in matching but also to its remarkable interpretability. The minutiae expressed by their origin (i.e. and coordinates) and angle are extracted first and then minutiae correspondences in the input fingerprint pair could be established and visualized. Similar to minutiae matching, Cao et al. [5] introduce the concept of virtual minutiae using deep learning models to sample directed points located on fingerprint images to ensure an adequate number of key points within the latent fingerprint area and to provide interpretable correspondences. Gu et al. [16] propose a coarse-to-fine matching scheme considering using undirected sampling points as key points to produce interpretable correspondences on the given latent fingerprint pair. However, while identified key points e.g. minutiae or virtual minutiae could be used to obtain the average similarity employed for fingerprint matching, they lack the benefit of the fixed-length representation extracted by deep models. Performing pairwise comparisons of fingerprints is computationally demanding which varies with the number of detected key points. On the other hand, Cui et al. [31] also define the task of fingerprint dense registration for obtaining pixel-wise displacement measures between two fingerprints with nonlinear skin distortion. The learned correspondences could contribute to the improvement of the subsequent matching performance as well as providing interpretable pixel-wise correspondences of feature points.
Despite the aforementioned studies being able to provide interpretable pixel-wise correspondences of feature points, none of them is integrated with learning fixed-length fingerprint representation during the matching process, which limits the reduction of computational complexity and applicability in other downstream tasks e.g. encryption [3]. In addition, the CNNs employed in these studies have limited receptive fields, which is not conducive to finding correspondences of feature points within a large global context [32]. A network called AFRNet is devised by Grosz et al. [17] for learning discriminative fixed-length fingerprint representations in the indexing task while obtaining correspondences between local fixed-length fingerprint representations that can be used for refining global representations in low-certainty situations. However, similar to previous studies, the CNN branch of AFRNet constrains the achievable number of correspondences, limiting its interpretability. On the other hand, the correspondences are patch-based instead of pixel-wise and are not a direct output of the neural network. They are obtained using a brute-force algorithm that demands significant computational resources, making the process time-consuming.
Employing gradients of the prediction output to identify parts of an input image that exert the most influence on the fingerprint matching result is also considered. Chowdhury et al. [33] train a multi-scale dilated Siamese CNN using fingerprint patches and demonstrate the significance of minutiae for fingerprint matching using Grad-CAM [34]. However, it could only reveal general areas that CNN focuses on during the inference stage, but could not establish further detailed pixel-wise correspondences.
In contrast to previous studies, we propose IFViT for fingerprint matching with excellent matching performance and interpretability. It combines the benefit of fixed-length representations extracted by deep models and interpretable pixel-wise correspondences of feature points in matching results. The ViT-based model is established to obtain interpretable dense pixel-wise matches. These matches could be employed for effective fingerprint alignment and improve interpretability in the subsequent matching procedure, allowing feature points used for fixed-length representation-based fingerprint matching to be visualized.
III Interpretable Fixed-Length Representation for Fingerprint Matching
In this section, we provide a detailed description of the proposed IFViT. Firstly, we provide a high-level overview and intuition of IFViT, then discuss the implementation mechanism of the ViT-based module for interpretable fingerprint dense registration, followed by the working mechanism of the ViT-based module for interpretable fingerprint fixed-length representation extraction and matching.
III-A Overview
The architecture of the proposed IFViT is shown in Fig. 2. It contains two main modules, an interpretable dense registration module and an interpretable fixed-length representation extraction and matching module. IFViT is trained with a large dataset composed of millions of fingerprint images across several benchmark datasets including plain, slap and rolled fingerprints.
Given the input fingerprint pair and , a specific CNN, ResNet-18 [35], is employed to extract local features and . The ViT is then adopted to learn the position and context-dependent features from and , followed by a differentiable matching layer [36] to produce dense pixel-wise correspondences of feature points for the given fingerprint pair. Alignment is subsequently carried out based on these correspondences to obtain an aligned fingerprint image with reference to the other input . The aligned fingerprint pair then undergoes fingerprint enhancement implemented by FingerNet [37], a state-of-the-art minutiae extractor consisting of orientation estimation, segmentation, Gabor enhancement and minutiae extraction, to obtain the enhanced fingerprint pair and . To take full advantage of the complementary local and global representation of the input fingerprint pair, we identify the corresponding Region of Interests (ROIs), including and in original fingerprint images, and the overlapped regions and in the aligned enhanced fingerprint images. Subsequently, those ROIs are passed into the ViT that is trained in the interpretable dense registration module but is finetuned with the new fully connected layer to obtain interpretable dense pixel-wise correspondences of feature points and discriminative fixed-length fingerprint representations, simultaneously. The learned local and global fixed-length representations are individually computed for their similarity, and combined to derive the final matching score.
III-B Interpretable Dense Registration
Dense fingerprint registration seeks to establish pixel-wise correspondences of feature points between two fingerprints that are subjected to nonlinear skin distortion or displacement [38]. These learned correspondences could then be used to fit a spatial transformation model for aligning fingerprints which is an essential component in most fingerprint recognition systems [3]. Since the quality of the alignment may greatly affect the performance of subsequent fingerprint matching process, it motivates us to develop a robust dense registration model. We particularly consider the characteristics of the ViT due to its ability to capture global context.
Given the pair of fingerprint images and , firstly we use a CNN to learn translation-equivariant local features which also reduces the input length to the ViT module to ensure a manageable computation cost. In this study, the ResNet-18 is chosen as the backbone. After obtaining feature maps from the CNN, a ViT model is built to extract local features that are both position-dependent and contextually relevant. Specifically, ViT performs a linear projection at first based on patches in feature maps and then transforms them into 1D tokens as shown below.
(1) |
where represents the linear projection. The position embedding is also added to tokens to provide the positional information. Subsequently, the token is passed into the encoder consisting of a sequence of transformer layers. As a critical part in the encoder, the attention layer receives the token in three different vectors: query (), key () and value (). The vectors originating from diverse inputs are subsequently consolidated into three different matrices, namely , and . The query retrieves information from the value based on the attention weights calculated from the dot product of (corresponding to each value ) and the key vector . The process is denoted as:
(2) |
In addition to the self-attention layer, the cross-attention layer inspired by [39, 32] is also introduced in our ViT model to enable cross-image communication in the fingerprint pair similar to the way humans look back-and-forth when matching images. In terms of the self-attention layer, the input features and are the same (either or ). However, for the cross-attention layer, the and are either ( and ) or ( and ). The self-attention layer and cross-attention layer are interleaved in our transformer by times. After passing through the encoder layer, the transformed features denoted by and are obtained.
In terms of identifying the dense pixel-wise correspondences of feature points in the given fingerprint pair, a differentiable matching layer using the dual-softmax operator is adopted [36]. Firstly, we compare the feature similarity for each pixel in with respect to all pixels in by calculating their correlations as follows:
(3) |
where correlation matrix denotes the correlation value of each coordinate between and . is a normalization factor to prevent the dot-product operation from producing large values.
Subsequently, we normalize using the softmax operation to obtain the matching confidence matrix as in equation (4). The utilization of a softmax-based approach not only enables the end-to-end training process but also provides pixel-level accuracy.
(4) |
The candidate matches in are selected if the corresponding confidence score is higher than a given threshold . We use Mutual-Nearest-Neighbor (MNN) criteria to further filter outlier matches and obtain the final selected matches :
(5) |
Loss function. For the interpretable dense registration module, the loss function for training the ViT that produces the dense pixel matches is negative log-likelihood loss, which is computed based on the matching confidence matrix . The pre-trained ViT model for feature matching in LoFTR [32] is employed to accelerate the convergence for training. Considering there is no available public fingerprint database that provides ground truth labels of matches of fingerprint pairs, we construct several noise models to simulate diverse types of noises and corruptions appearing in the real fingerprint images to improve the model’s ability for recognizing correspondences of feature points in low-quality or low-texture fingerprint images. In this case, the fingerprint images enhanced by FingerNet are not used but instead, we solely rely on original fingerprint images and their corresponding corrupted image to train the model. Since the relation of each pixel position between original and synthetic corrupted fingerprint images could be directly obtained, we are able to determine ground-truth labels for the confidence matrix. Three types of noises are simulated in this study including sensor noise, dryness and over-pressurization, as depicted in Fig. 3. Sensor noise could be approximated by the Perlin noise [40], whilst dryness and over-pressurization could be simulated by performing the dilation and erosion operation, respectively. To ensure rotation invariance, all corrupted fingerprint images are rotated randomly within 60 degrees. This rotation parameter is determined based on the domain knowledge regarding the maximum degree to which a user would rotate their fingers while positioning them on the reader platen [3]. Furthermore, to minimize incorrect matches, ground-truth labels are also created for the registration of imposter fingerprint pairs. The loss function of dense registration based on negative log-likelihood loss over the grids in is minimized during the training stage:
(6) |

III-C Interpretable Fixed-length Representation Extraction and Matching
After performing the dense registration on the given fingerprint pair, we establish the second critical component of our method, namely the interpretable fixed-length-representation extraction-and-matching module. In this module, we combine both local and global representations of fingerprints to improve the matching performance. It can additionally provide interpretable dense pixel-wise correspondences of feature points similar to the previous interpretable dense registration model in the matching results.
Fingerprint Alignment. Recognizing that the outcome of the alignment would significantly influence the subsequent matching module’s performance, we rotate at five angles (-60°, -30°, 0°, 30°, 60°) to determine the best matches for alignment. Subsequently, we compute the homography matrix using the RANSAC algorithm and apply the obtained directly to to produce the aligned . Compared with the Spatial Transformer Network (STN) which has been widely used recently for fingerprint alignment, e.g., in [3, 6], our proposed ViT-based interpretable dense registration module for fingerprint alignment can provide dense pixel-wise correspondences of feature points for the given fingerprint pair, which is conducive to understanding the criteria used to determine when fingerprint alignment succeeds or fails.
Obtaining Local and Global Representation. Although the global fixed-length fingerprint representation has been demonstrated to be effective for fingerprint recognition [3, 41, 42], it may produce high similarity on some global similar ridge-flow structures from different identities by mistake, therefore leading to wrong matches [19]. On the other hand, local representation, e.g. fingerprint patches, may also face matching failures, primarily because some factors like distortion and dryness could negatively impact the effective fingerprint information, especially for minutiae [19]. Fig. 4 shows a case susceptible to matching failures when considering either only the local or global representation. Therefore, combinations of both representations should be considered to compensate for the semantic shortcomings of each other.

In this study, we fuse global representation extracted from overlapped regions of enhanced fingerprints with local representation derived from the corresponding ROI of original fingerprints. The motivation for considering the ROI from original fingerprints instead of enhanced fingerprints as the local representation is that we observed a certain level of prediction error when using FingerNet for enhancement. The FingerNet performs well in enhancing the global ridge-flow structures of fingerprints but is prone to introducing patch-like artifacts, resembling a mosaic, in low-quality local areas in fingerprints. This could potentially lead to the loss of some crucial information.
Specifically, we introduce the Sobel operator to calculate the gradient of the fingerprint image’s intensity in the horizontal () and vertical () directions, and then compute the magnitude of the gradient for each pixel based on equation (7). The box filter with 25×25 kernel size is used to obtain the integral of the gradient magnitude. Threshold which equals 15 percent of the maximum integral of the obtained gradient magnitude is further empirically determined for segmenting the general fingerprint area. We perform detection for contours on the segmented regions and conduct connected component analysis, resulting in the respective mask of an effective fingerprint area for and . After taking the intersection of masks, they are separately applied to and , producing effective overlapped fingerprint areas and (i.e. global representation). The flowchart for extracting overlapped regions from the fingerprint pair is illustrated in Fig. 5. Although the FingerNet can also directly output the mask of a fingerprint, the generated mask frequently includes irrelevant background around the fingerprint, as shown in and of Fig. 5. For the model to focus only on the effective fingerprint information, the mask generated by FingerNet is not considered in this step. Furthermore, we extract ROIs (90×90) from the overlapped region of original fingerprints as the local representation.
(7) |
Learning Interpretable Fixed-length Representation. For the interpretable fixed-length representation extraction and matching module, we propose two ViT-based Siamese Networks with shared weights which learns a similarity metric for the input fingerprint pair and outputs a matching score. Specifically, the input consists of the overlapped fingerprint and (i.e. global information) as well as the corresponding ROIs and extracted from original fingerprint images (i.e. local information). We utilize the interpretable dense registration module trained previously in the fingerprint matching module to generate pseudo labels of fingerprint pairs for learning correspondences of feature points. We combine the previously trained ViT-based Siamese model as the backbone with a new fully connected layer, which is then retrained with labelled fingerprint pairs. Thus, it offers interpretable dense pixel-wise correspondences of feature points in matching results, and the 256-dimensional fixed-length representations and .

Loss Function. In this module, dense registration loss is employed again to learn dense pixel-wise correspondences of feature points for the fingerprint pair from ViT’s output. The cosine embedding loss shown in equation (8) is chosen for training the model to transform identified correspondences of feature points into the 256-dimensional discriminative fixed-length representation.
(8) |
To minimize and maximize the distance of learned intra-class and inter-class representation separately, we further introduce the ArcFace loss in equation (9) to optimize the network via classification task, constituting the total loss shown in equation (10).
(9) |
(10) |
where represents the ground truth of the cosine embedding loss and is the true class label of the ArcFace loss , respectively. is the total number of the training samples while is the number of classes. denotes the cosine of the angle between the learned representation ( or ) and the weight vector for the corresponding label . In terms of , it calculates the cosine of the angle between the learned representation ( or ) and the weight vector for all classes except the . Additionally, the margin is set to 0.4 for both and in this study and the rescaling factor is set as 64. , and in the are empirically set as 0.5, 0.1 and 1 to balance the matching performance and interpretability, respectively.
Rather than introducing the domain knowledge (i.e. minutiae information) [6, 4, 3], we hope that the fixed-length representation learned from the IFViT could utilize prior information related to dense pixel-wise correspondences of feature points in the fingerprint pair, to deal with the situation where it is challenging to learn minutiae information from low-quality fingerprints.
Matching Score Computation. For evaluation, the learned fixed-length global representation , and fixed-length local representation , of an input fingerprint pair can be extracted. The matching scores, namely and based on the , and , are calculated as shown in equation (11) and normalized to the range [-1, 1]. The total matching score is finally obtained by weighted fusion of the individual scores ( and ) as shown in equation (12), where the weights and are determined by linear regression learning.
(11) |
(12) |
IV Experiments
In this section, we first introduce several public fingerprint datasets employed in this study along with a detailed experimental setup. Then we conduct the performance evaluation of the interpretable dense registration module and the interpretable fixed-length representation extraction and matching module in the proposed IFViT.
IV-A Datasets
Dense Registration Module. Regarding the training of the interpretable dense registration module, the FVC2002 (DB1, DB2 and DB3) [43], NIST SD301a (A, B, C, E, J, K M and N) [44], a subset of NIST SD302a (A, B, C, D, E, F, U, V, L and M) [45] and a subset of MOLF (DB1 and DB2) [46] are used. The total number of fingerprint images in these datasets is 25, 090, and after applying three noise models previously proposed, a total of 100, 360 training images can be obtained. They are combined to form 100K fingerprint pairs (75K for genuine pairs and 25K for imposter pairs) for training.
Matching Module. In terms of the interpretable fixed-length representation extraction and matching module, we employ FVC2002 (DB1, DB2 and DB3) [43], NIST SD301a (A, B, C, E, J, K M and N) [44], a subset of NIST SD302a (A, B, C, D, E, F, U, V, L and M) [45], a subset of MOLF (DB1 and DB2) [46] and a recently released synthetic fingerprint dataset called PrintsGAN [47] as the training dataset and evaluate the model on FVC2004 (DB1, DB2 and DB3) [48], NIST SD4 [49], the remaining test partitions of NIST SD302a (last 200 fingers) and MOLF (last 200 fingers). There is no overlap between fingers selected for training and testing sets in NIST SD302a and MOLF. Furthermore, to strike a balance between model performance and computational demand, we only use the first impression from different fingers when creating imposter fingerprint pairs; otherwise, the size of the training dataset could even reach 100 M. Additionally, in PrintsGAN, only 2, 500 fingers are selected as a part of our training set and mainly used for model pre-training. In total, our aggregated training dataset contains 3.6M pairs (7.2M images). Table I provides the details of all databases and their usage in this study.
Database
FVC2002 DB1A [43]
FVC2002 DB2A [43]
FVC2002 DB3A [43]
FVC2004 DB1A [48]
FVC2004 DB2A [48]
FVC2004 DB3A [48]
Example Image
Fingers Used
100
100
100
100
100
100
Images Used
800
800
800
800
800
800
Usage
Training (7, 750 pairs)
Training (7, 750 pairs)
Training (7, 750 pairs)
Testing (7, 750 pairs)
Testing (7, 750 pairs)
Testing (7, 750 pairs)
Database
NIST SD301a [44]
NIST SD302a [45]
NIST SD4 [49]
MOLF [46]
PrintsGAN [47]
Example Image
Fingers Used
240
1, 789
2, 000
1, 000
2500
Images Used
1, 920
17, 890
4, 000
8, 000
37, 500
Usage
Training (9, 480 pairs)
Training (193, 047 pairs)
Validation (18, 316 pairs)
Testing (28, 900 pairs)
Testing (251, 000 pairs)
Training (62, 000 pairs)
Testing (15, 500 pairs)
Training (3, 386, 250 pairs)
Method MI of ROIs Extracted from the Original Fingerprint Pair ( and ) MI of the Enhanced Fingerprint Pair ( and ) FVC 2004 DB1 FVC 2004 DB2 FVC 2004 DB3 MOLF NIST SD4 NIST SD302a FVC 2004 DB1 FVC 2004 DB2 FVC 2004 DB3 MOLF NIST SD4 NIST SD302a Unaligned 0.42 0.19 0.40 0.26 0.15 0.36 0.47 0.49 0.69 0.56 0.42 0.37 Dense Registration in IFViT 0.54 0.38 0.46 0.33 0.19 0.35 0.74 0.90 0.90 0.91 0.83 0.74 SIFT 0.24 0.33 0.34 0.14 0.16 0.11 0.27 0.31 0.31 0.28 0.22 0.19 STN_DeepPrint 0.70 0.19 0.40 0.30 0.15 0.34 0.65 0.87 0.85 0.85 0.78 0.72 STN_AFRNet 0.71 0.18 0.40 0.32 0.14 0.38 0.64 0.87 0.84 0.83 0.78 0.68
Method EER of ROIs Extracted from the Original Fingerprint Pair ( and ) EER of the Overlapped Region Extracted from the Enhanced Fingerprint Pair ( and ) FVC 2004 DB1 FVC 2004 DB2 FVC 2004 DB3 MOLF NIST SD4 NIST SD302a FVC 2004 DB1 FVC 2004 DB2 FVC 2004 DB3 MOLF NIST SD4 NIST SD302a Unaligned 7.75% 7.38% 6.88% 6.42% 7.35% 5.49% 5.52% 5.32% 5.15% 7.47% 6.82% 7.00% Dense Registration in IFViT 2.87% 2.40% 2.57% 2.93% 3.58% 3.12% 2.71% 2.60% 2.67% 3.37% 4.31% 2.85% SIFT 27.42% 26.53% 28.77% 19.33% 42.03% 23.42% 23.31% 16.55% 39.66% 31.75% 44.22% 25.62% STN_DeepPrint 4.77% 4.72% 4.75% 5.79% 6.63% 5.27% 4.88% 4.83% 4.82% 5.77% 6.44% 5.43% STN_AFRNet 4.31% 4.22% 4.27% 5.28% 6.82% 5.06% 4.34% 4.22% 4.23% 5.92% 6.45% 5.08%
IV-B Implementation Details
In this study, the interpretable dense registration module and the interpretable fixed-length representation extraction and matching module in IFViT are implemented based on Pytorch and trained independently on four NVIDIA GeForce RTX 3090 GPUs.
In terms of the interpretable dense registration module, a total of 200K fingerprint images (100K fingerprint pair) are used for training. The input images are resized to 128 × 128 to accelerate the convergence process. The module is trained over 100 epochs with a learning rate of . For the interpretable fixed-length representation extraction and matching module, we initially pretrain it employing 6.7M synthetic images from PrintsGAN with a learning rate of . Then, 575K real fingerprint images are used for fine-tuning this module over 70 epochs with a learning rate of . Both of these two modules are optimized by Adam with a weight decay of and a batch size of 128.
IV-C Performance of Fingerprint Dense Registration
Due to the lack of ground truth labels for pixel-wise correspondences of feature points in real fingerprint pairs, it is challenging to evaluate the performance of the fingerprint dense registration algorithm directly through the positions of correspondences. Instead, we execute alignment using correspondences of feature points produced from the registration algorithm and evaluate the alignment performance based on the similarity of aligned enhanced fingerprint pairs ( and ) and corresponding ROIs ( and ) in original fingerprint pairs. Notably, we do not select overlapped regions from enhanced fingerprint pairs ( and ) for similarity evaluation due to the potential risk of alignment failure, which could result in an extremely limited overlap region. It might lead to similarity computations being based on invalid background areas, finally leading to unreasonable similarity outcomes.
We select the Mutual Information (MI) metric that has been widely used for image registration in the last decades [50, 51] to measure the similarity in aligned fingerprint pairs. Higher MI observed in the fingerprint pair from the same finger indicates superior performance by the registration algorithm. In this study, we compare the alignment performance of the proposed module with two methods: a prevalent open-source algorithm, Scale-Invariant Feature Transform (SIFT), employed for recent registration in fingerprint matching [33], and the Spatial Transformer Network (STN), an alignment module widely adopted in many fingerprint recognition tasks [3, 17, 6]. The STNs trained in DeepPrint and AFRNet [3, 17] are introduced due to their state-of-the-art performance on fingerprint benchmark datasets.


We report MI values based on different methods in Table II. As we can observe 1) The performance of the interpretable dense registration module in the proposed IFViT is more accurate than other methods with better alignment performance. This is evident in its highest MI value on ROIs of original fingerprint pairs and aligned enhanced fingerprint pairs from the same finger across most datasets. 2) The MI values on aligned enhanced fingerprint pairs are generally higher than those on ROIs of original fingerprint pairs. This is primarily attributed to the MI value being influenced by many factors e.g. the quality disparities in original fingerprint pairs, and the variations in fingerprint appearance caused by various sensors. However, following enhancement by FingerNet, all enhanced fingerprint images adhere to uniform quality standards, reducing their inconsistencies. In this context, the MI value of the latter provides a more precise indication of fingerprint alignment performance. 3) Relative to MI values of unaligned fingerprint pairs, MI values of aligned fingerprint pairs using the SIFT algorithm have shown a marked decline. This is due to the propensity of SIFT to generate incorrect correspondences of feature points, which in turn result in erroneous affine transformations during the alignment procedure. It compromises the structural integrity of original fingerprint images, as illustrated in Fig. 6 (b). 4) The STN in AFRNet and DeepPrint show a comparable level of alignment performance. Notably, STN in AFRNet has superior MI values on ROIs of original fingerprint pairs in the FVC2004 DB1 and NIST SD302a. However, it has slightly worse MI values for enhanced fingerprint pairs compared to the proposed dense registration module. Besides, due to the limitation of STNs in these models to only perform rotation and translation, they cannot achieve pixel-wise alignment for cross-sensor fingerprint pairs, as shown in Fig. 6 (c) (d).
We further explore the influence of fingerprint alignment from various methods on the matching performance based on the trained proposed model, and report their Equal Error Rate (EER) in Table III. The results indicate that the quality of fingerprint alignment significantly impacts the model’s matching performance. When considering pixel-wise correspondences of feature points established by the proposed dense registration module, the aligned fingerprints demonstrate the lowest EER in the ROIs from original fingerprint pairs and the overlapped region extracted from aligned enhanced fingerprint pairs across diverse datasets. In comparison to the other methods, our approach is more robust, even for cross-sensor fingerprint pairs, resulting in a considerable number of precise correspondences of feature points between them as shown in Fig. 6 (a).
IV-D Performance of Fingerprint Matching
We compare the matching performance of the proposed IFViT with two recent well-known works in fingerprint recognition called DeepPrint and AFRNet [3, 17]. Aside from the minutiae points in the DeepPrint network, which are generated by FingerNet as an alternative for creating minutiae maps, the employed model structures are as consistent as possible with those reported in their studies.
From the EER reported in Table IV and the full Receiver Operating Characteristic (ROC) shown in Fig. 7, it can be observed that the proposed IFViT outperforms the other two approaches across six benchmark databases, achieving the lowest EER. The reason for this discrepancy is that we observe a decrease in the performance of DeepPrint and AFRNet when there is an imbalance in the quantity of fingerprints collected from different manners (e.g., rolled, plain, and slap) on the training set. In the context of a limited number of fingers, the developed IFViT can consider both the ROIs of the original fingerprint pairs and the overlapped regions from enhanced fingerprint pairs to reduce the disparity among fingerprint images, thus improving the model’s generalization capacity. Particularly with cross-sensor fingerprints from the MOLF and NIST SD302a, predicting them poses a challenge for both DeepPrint and AFRNet. But for IFViT, it can notably reduce the EER by integrating the dense registration module for pixel-wise precise alignment and combining local and global representations within the matching module.
Method EER FVC 2004 DB1 FVC 2004 DB2 FVC 2004 DB3 MOLF NIST SD4 NIST SD302a IFViT 1.87% 1.83% 1.49% 2.33% 2.71% 2.31% DeepPrint 2.70% 3.39% 2.05% 5.18% 2.85% 5.03% AFRNet 2.43% 2.95% 1.56% 5.33% 2.80% 6.29%

IV-E Interpretability
In addition to demonstrating the interpretability of fingerprint alignment shown in Fig. 6. We further investigate the interpretability of the proposed matching module. Based on the implementation reported by AFRNet [17], we evaluate the interpretability of AFRNet by computing the correspondences of feature points in the fingerprint pairs. In addition, similar operations are also performed on the CNN output of DeepPrint [3] to compare with our method on original fingerprint images. As shown in Fig. 8, when matching fingerprint pairs from the same identity, the matching module in the proposed IFViT can offer more reasonable and superior interpretability. DeepPrint can only focus on local regions and establish limited number of correspondences. AFRNet, due to its adoption of self-attention in ViT that can capture long-range dependencies and global context, similar to the proposed IFViT, can provide more correspondences in the global context, but it is challenging to provide reasonable ones based on the matching result.
Fig. 9 further illustrates the interpretability of IFViT during the matching procedure by using the fusion of local and global representations. It can calculate the matching score for the given fingerprint pair while simultaneously establishing interpretable dense pixel-wise correspondences of feature points across both the ROI extracted from the original fingerprint ( and ) and the overlapped region from the enhanced fingerprint ( and ) in the matching result. This aids in understanding the reasons behind the decision made by the deep learning model during fingerprint matching. In the example provided in Fig. 9, the imposter pair establishes fewer correspondences of feature points during prediction compared to the genuine pair, with the vast majority of them being incorrect.

To further observe which feature points of fingerprints are possibly used in fixed-length representation-based matching, we provide examples of learned correspondences with a high confidence level. Fig. 10 (a) showcases that feature points of fingerprints at different levels can be identified around learned dense correspondences by IFViT. These features range from the global-level feature (level-1) such as the loop, to local-level features (level-2) represented by minutiae composed of ridge endings and bifurcations, down to the very fine-level feature (level-3) typified by the pore. This implies that, although we do not explicitly integrate domain knowledge into the model, to some extent, the model can infer it on its own. Moreover, we also discover that as the resolution of fingerprint images decreases, such as in cases where pores cannot be clearly distinguished, or when the matching region is limited, there is a noticeable shift towards emphasizing the minutiae in learned dense correspondences as shown in Fig. 10 (b). This is aligned with our understanding that minutiae are one of the most important features adopted in fingerprint matching algorithms.
Instead of using techniques such as Grad-CAM or Saliency Map [33, 6], which rely on gradient computation to identify general regions contributing to predictions in the input fingerprint pair, IFViT can provide superior interpretability at the pixel level. It allows for the detailed examination of specific fingerprint feature points that are used and transformed into a fixed-length representation for matching, offering insights into the decision-making process of black-box deep learning models for both successful and failed fingerprint matching cases.

IV-F Ablation Study
Method Alignment Local Representation Global Representation Fusion of Local and Global Representation EER FVC 2004 DB1 FVC 2004 DB2 FVC 2004 DB3 MOLF NIST SD4 NIST SD302a Proposed Method (A) ✓ 47.63% 46.40% 47.07% 36.80% 35.36% 51.63% Proposed Method (B) ✓ 31.16% 30.14% 29.66% 42.52% 12.92% 39.69% Proposed Method (C) ✓ ✓ 2.87% 2.40% 2.57% 2.93% 3.58% 3.12% Proposed Method (D) ✓ ✓ 2.71% 2.60% 2.67% 3.37% 4.31% 2.85% Proposed Method (E) ✓ ✓ ✓ ✓ 1.87% 1.83% 1.49% 2.33% 2.71% 2.31%
To justify the incorporation of each component (i.e. fingerprint alignment, local representation and global representation), we conduct an ablation study by sequentially introducing each module into the matching pipeline. The outcomes, as shown in Table V, demonstrate improved matching performance with the addition of each component. It highlights the importance of the proposed alignment module in improving fingerprint matching performance, especially in cases involving fingerprint pairs with limited overlapping areas. Furthermore, the combination of local and global representations can compensate for the deficiencies in semantic information present in each, thereby achieving an improved recognition rate.
IV-G Computational Efficiency
The inference time with the proposed IFViT primarily involves three stages: fingerprint enhancement by FingerNet, registration and alignment, and matching. Running on an Intel(R) Xeon(R) Gold 6133 CPU @ 2.50GHz with the NVIDIA GeForce RTX 3090 GPU, it takes approximately 463ms for inference based on a given fingerprint pair (116ms for enhancement, 187ms for registration and alignment, and 160ms for matching). To further improve the computational efficiency, the deep convolutional neural network (i.e. ResNet-18) for local feature extraction in IFViT could be replaced by other shallower networks. Also, the FingerNet could be substituted with alternative fingerprint enhancement techniques with lower computational complexity. We do not consider optimization of them, as this falls outside the scope of this study.
V Conclusion
In this paper, we propose a novel multi-stage interpretable fingerprint matching network via vision transformer called IFViT. This is the first attempt to utilize the ViT to achieve fingerprint dense registration. The proposed interpretable fingerprint registration module can generate dense pixel-wise correspondences of feature points and perform precise alignment for the input fingerprint pair. The interpretable fixed-length representation extraction and matching module, achieves lower EER and higher recognition accuracy across multiple datasets when the model is trained on a limited number of fingers, simultaneously demonstrating superior interpretability. In our future work, we plan to further optimize the existing framework and explore its applicability in more challenging scenarios e.g. contactless fingerprints and latent fingerprints to enhance their interpretability during the matching process.
Acknowledgment
The authors are grateful to Zhiyu Pan from Tsinghua University for his valuable suggestions and insights that helped us to improve the quality of this work.
References
- [1] M. Chhabra, M. K. Shukla, and K. K. Ravulakollu, “State-of-the-art: A systematic literature review of image segmentation in latent fingerprint forensics,” Recent Advances in Computer Science and Communications, vol. 13, no. 6, pp. 1115–1125, 2020.
- [2] S. Arora and M. S. Bhatia, “Fingerprint spoofing detection to improve customer security in mobile financial applications using deep learning,” Arabian Journal for Science and Engineering, vol. 45, no. 4, pp. 2847–2863, 2020.
- [3] J. J. Engelsma, K. Cao, and A. K. Jain, “Learning a fixed-length fingerprint representation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 43, no. 6, pp. 1981–1997, 2019.
- [4] S. A. Grosz, J. J. Engelsma, E. Liu, and A. K. Jain, “C2cl: Contact to contactless fingerprint matching,” IEEE Transactions on Information Forensics and Security, vol. 17, pp. 196–210, 2021.
- [5] K. Cao, D. L. Nguyen, C. Tymoszek, and A. K. Jain, “End-to-end latent fingerprint search,” IEEE Transactions on Information Forensics and Security, vol. 15, pp. 880–894, 2019.
- [6] A. Takahashi, Y. Koda, K. Ito, and et al., “Fingerprint feature extraction by combining texture, minutiae, and frequency spectrum using multi-task cnn,” in 2020 IEEE International Joint Conference on Biometrics (IJCB). IEEE, 2020, pp. 1–8.
- [7] Z. Lin, G. Zhu, Y. Deng, X. Chen, Y. Gao, K. Huang, and Y. Fang, “Efficient parallel split learning over resource-constrained wireless edge networks,” IEEE Transactions on Mobile Computing, 2023.
- [8] Y. Qiu, Y. Hui, P. Zhao, C. H. Cai, B. Dai, J. Dou, et al., “A novel image expression-driven modeling strategy for coke quality prediction in the smart cokemaking process,” Energy, vol. 294, p. 130866, 2024.
- [9] H. Yuan, Z. Chen, Z. Lin, J. Peng, Z. Fang, Y. Zhong, Z. Song, X. Wang, and Y. Gao, “Graph Learning for Multi-Satellite Based Spectrum Sensing,” in Proc. ICCT, 2023.
- [10] Z. Lin, G. Qu, W. Wei, X. Chen, and K. K. Leung, “Adaptsfl: Adaptive split federated learning in resource-constrained edge networks,” arXiv preprint arXiv:2403.13101, 2024.
- [11] Y. Qiu, J. Wang, and et al., “Pose-guided matching based on deep learning for assessing quality of action on rehabilitation training,” Biomedical Signal Processing and Control, vol. 72, p. 103323, 2022.
- [12] R. Confalonieri, L. Coba, and et al., “A historical perspective of explainable artificial intelligence,” Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery, vol. 11, no. 1, p. e1391, 2021.
- [13] F. K. Došilović, M. Brčić, and N. Hlupić, “Explainable artificial intelligence: A survey,” in 2018 41st International convention on information and communication technology, electronics and microelectronics (MIPRO). IEEE, May 2018, pp. 0210–0215.
- [14] M. Tico and P. Kuosmanen, “Fingerprint matching using an orientation-based minutia descriptor,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, no. 8, pp. 1009–1014, 2003.
- [15] Z. Cui, J. Feng, and J. Zhou, “Dense registration and mosaicking of fingerprints by training an end-to-end network,” IEEE Transactions on Information Forensics and Security, vol. 16, pp. 627–642, 2020.
- [16] S. Gu, J. Feng, J. Lu, and J. Zhou, “Latent fingerprint registration via matching densely sampled points,” IEEE Transactions on Information Forensics and Security, vol. 16, pp. 1231–1244, 2020.
- [17] S. A. Grosz and A. K. Jain, “Afr-net: Attention-driven fingerprint recognition network,” IEEE Transactions on Biometrics, Behavior, and Identity Science, 2023.
- [18] C. Lin and A. Kumar, “A cnn-based framework for comparison of contactless to contact-based fingerprints,” IEEE Transactions on Information Forensics and Security, vol. 14, no. 3, pp. 662–676, 2018.
- [19] S. Tandon and A. Namboodiri, “Transformer based fingerprint feature extraction,” in 2022 26th International Conference on Pattern Recognition (ICPR). IEEE, August 2022, pp. 870–876.
- [20] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, and N. Houlsby, “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
- [21] J. Chen and C. M. Ho, “Mm-vit: Multi-modal video transformer for compressed video action recognition,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2022, pp. 1910–1921.
- [22] Y. Huang, L. Chen, C. Zhou, and et al., “Model long-range dependencies for multi-modality and multi-view retinopathy diagnosis through transformers,” Knowledge-Based Systems, vol. 271, p. 110544, 2023.
- [23] J. Yang, C. Li, P. Zhang, and et al., “Focal attention for long-range interactions in vision transformers,” Advances in Neural Information Processing Systems, vol. 34, pp. 30 008–30 022, 2021.
- [24] K. Cao and A. K. Jain, “Fingerprint indexing and matching: An integrated approach,” in 2017 IEEE International Joint Conference on Biometrics (IJCB). IEEE, October 2017, pp. 437–445.
- [25] S. A. Grosz, J. J. Engelsma, R. Ranjan, N. Ramakrishnan, M. Aggarwal, G. G. Medioni, and A. K. Jain, “Minutiae-guided fingerprint embeddings via vision transformers,” arXiv preprint arXiv:2210.13994, 2022.
- [26] W. Samek and K. R. Müller, “Towards explainable artificial intelligence,” Explainable AI: interpreting, explaining and visualizing deep learning, pp. 5–22, 2019.
- [27] P. C. Neto and et al., “Pic-score: Probabilistic interpretable comparison score for optimal matching confidence in single-and multi-biometric face recognition,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 1021–1029.
- [28] B. Yin, L. Tran, H. Li, X. Shen, and X. Liu, “Towards interpretable face recognition,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 9348–9357.
- [29] X. Jiang and W. Y. Yau, “Fingerprint minutiae matching based on the local and global structures,” in Proceedings 15th International Conference on Pattern Recognition. ICPR-2000, vol. 2. IEEE, September 2000, pp. 1038–1041.
- [30] R. Cappelli, M. Ferrara, and D. Maltoni, “Minutia cylinder-code: A new representation and matching technique for fingerprint recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 32, no. 12, pp. 2128–2141, 2010.
- [31] Z. Cui, J. Feng, and J. Zhou, “Dense fingerprint registration via displacement regression network,” in 2019 International Conference on Biometrics (ICB). IEEE, June 2019, pp. 1–8.
- [32] J. Sun, Z. Shen, Y. Wang, H. Bao, and X. Zhou, “Loftr: Detector-free local feature matching with transformers,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 8922–8931.
- [33] A. Chowdhury, S. Kirchgasser, A. Uhl, and A. Ross, “Can a cnn automatically learn the significance of minutiae points for fingerprint matching?” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2020, pp. 351–359.
- [34] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, and D. Batra, “Grad-cam: Visual explanations from deep networks via gradient-based localization,” in Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 618–626.
- [35] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778.
- [36] M. Tyszkiewicz, P. Fua, and E. Trulls, “Disk: Learning local features with policy gradient,” Advances in Neural Information Processing Systems, vol. 33, pp. 14 254–14 265, 2020.
- [37] Y. Tang, F. Gao, and et al., “Fingernet: An unified deep network for fingerprint minutiae extraction,” in 2017 IEEE International Joint Conference on Biometrics (IJCB). IEEE, October 2017, pp. 108–116.
- [38] X. Si, J. Feng, B. Yuan, and J. Zhou, “Dense registration of fingerprints,” Pattern Recognition, vol. 63, pp. 87–101, 2017.
- [39] P. E. Sarlin and et al., “Superglue: Learning feature matching with graph neural networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 4938–4947.
- [40] R. Cappelli, D. Maio, and D. Maltoni, “An improved noise model for the generation of synthetic fingerprints,” in ICARCV 2004 8th Control, Automation, Robotics and Vision Conference, 2004, vol. 2. IEEE, December 2004, pp. 1250–1255.
- [41] D. Song and J. Feng, “Fingerprint indexing based on pyramid deep convolutional feature,” in 2017 IEEE International Joint Conference on Biometrics (IJCB). IEEE, October 2017, pp. 200–207.
- [42] J. J. Engelsma, D. Deb, K. Cao, and et al., “Infant-id: Fingerprints for global good,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 7, pp. 3543–3559, 2021.
- [43] “Fvc2002,” http://bias.csr.unibo.it/fvc2002/, 2002.
- [44] G. Fiumara, P. Flanagan, M. Schwarz, E. Tabassi, and C. Boehnen, NIST special database 301, Gaithersburg, MD, USA, 2018.
- [45] G. P. Fiumara, P. A. Flanagan, J. D. Grantham, K. Ko, K. Marshall, M. Schwarz, and C. Boehnen, NIST special database 302: Nail to nail fingerprint challenge, 2019.
- [46] A. Sankaran, M. Vatsa, and R. Singh, “Multisensor optical and latent fingerprint database,” IEEE Access, vol. 3, pp. 653–665, 2015.
- [47] J. J. Engelsma, S. Grosz, and A. K. Jain, “Printsgan: Synthetic fingerprint generator,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 5, pp. 6111–6124, 2022.
- [48] “Fvc2004,” http://bias.csr.unibo.it/fvc2004/, 2004.
- [49] C. I. Watson and C. L. Wilson, “Nist special database 4,” Fingerprint Database, National Institute of Standards and Technology, vol. 17, no. 77, p. 5, 1992.
- [50] F. Maes, D. Vandermeulen, and P. Suetens, “Medical image registration using mutual information,” Proceedings of the IEEE, vol. 91, no. 10, pp. 1699–1722, 2003.
- [51] F. Maes, A. Collignon, and et al., “Multimodality image registration by maximization of mutual information,” IEEE Transactions on Medical Imaging, vol. 16, no. 2, pp. 187–198, 1997.