This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Unsupervised Deep Learning Image Verification Method

Enoch Solomon Department of Computer Science
Virginia State University
Richmond, Virginia
[email protected]
   Abraham Woubie Silo AI
Helsinki, Finland
[email protected]
   Eyael Solomon Emiru Department of Information
Engineering and Computer Science
University of Trento, Italy
[email protected]
Abstract

Although deep learning are commonly employed for image recognition, usually huge amount of labeled training data is required, which may not always be readily available. This leads to a noticeable performance disparity when compared to state-of-the-art unsupervised face verification techniques. In this work, we propose a method to narrow this gap by leveraging an autoencoder to convert the face image vector into a novel representation. Notably, the autoencoder is trained to reconstruct neighboring face image vectors rather than the original input image vectors. These neighbor face image vectors are chosen through an unsupervised process based on the highest cosine scores with the training face image vectors. The proposed method achieves a relative improvement of 56% in terms of EER over the baseline system on Labeled Faces in the Wild (LFW) dataset. This has successfully narrowed down the performance gap between cosine and PLDA scoring systems.

Index Terms:
autoencoder, biometrics, deep learning, face recognition, image recognition
publicationid: pubid: 979-8-3503-9341-5/23/$31.00 ©2023 IEEE

I Introduction

Face recognition is a system used for identifying or verifying a person from an image or video [1, 2, 4, 5, 6, 7, 8, 3]. Deep learning techniques, as a front-end approach, have demonstrated proficiency in acquiring intricate features [9], including bottleneck features (BNF) [10]. These features are subsequently integrated into traditional frameworks or employed in the extraction of face image vectors for the face recognition process. Deep learning methods prove effective in compressing face images into embeddings, as exemplified in [11] in addition to different speech and security applications [12, 15, 13, 14, 16, 17, 18, 19, 20, 21, 22].

Several works have been proposed to enhance face image verification performance through unsupervised learning techniques [23, 24, 25]. For instance, [26] introduces a vector representation of face images using RBM adaptation. On the backend, [27] proposes various imposter selection algorithms aimed at narrowing the performance gap between supervised and unsupervised face recognition methods. They apply DBN adaptation as a backend for face images. These algorithms necessitate training a separate model for each target person, incurring computational costs. However, the results still fall short of the supervised approach, which benefits from actual face image labels.

In this work, we focus on mitigating the necessity for labeled face image data in face verification. Our objective is to narrow the performance disparity between supervised and unsupervised techniques, all without relying on face image labels. Unlike traditional DNN-based classifiers, the training of autoencoders is an unsupervised process that operates independent of labeled data. Our contribution lies in the novel training approach for the autoencoder, specifically designed to account for session variability among face images in the absence of labeled data. Rather than reconstructing the same training face image, we train the autoencoder to reconstruct neighboring face images. Following this training, we extract face image vector embeddings for testing purposes. During experimental trials, we apply cosine scoring to evaluate the face image embedding vectors, which have exhibited a remarkable discriminatory capability. The experimental findings demonstrate a substantial 56% enhancement in performance over the baseline system that employs the cosine scoring technique, underscoring the efficacy of our proposed autoencoder training methodology.

The rest of the paper is as follows: Section 2 provides an in-depth description of the proposed autoencoder training method and the process of neighbor face image selection. Section 3 describes the experimental configuration and outlines the dataset employed. Section 4 describes the experimental results. Finally, Section 5 gives the conclusion of the proposed work.

Refer to caption
Figure 1: Proposed approach for training the autoencoder.

II Proposed Method

In this work, we introduce an unsupervised deep learning method for face verification. The training procedure of the proposed method is outlined as follows: From a pool of unlabeled large-scale face image data, we identify the k nearest neighbor images that closely resemble the specified face image, which are then employed for training. Diverging from the conventional deep neural network training, our method computes the loss function for each input face image k times.

PLDA scoring technique for face image vectors necessitates labels [28], which can be challenging to obtain in practical scenarios. Conversely, the cosine scoring technique operates without the need for labels, albeit at the expense of performance degradation. To bridge the achievement gap between these two scoring methods in the absence of labeled training data, we introduce a novel framework for autoencoder training. This approach is entirely unsupervised, distinguishing it from conventional DNN classifiers and PLDA. Additionally, we advocate training the autoencoder to reconstruct similar face images, rather than exclusively focusing on the same training face image. Through this method, the autoencoder, trained in an unsupervised manner, effectively addresses session variability among face images without relying on labels.

II-A Autoencoder training

As shown in Fig 1, an autoencoder comprises an encoder and a decoder. The encoder functions to compress the input face image vector xx into a lower-dimensional space, while the decoder reconstructs it back to its original form. Conventional training involves minimizing the Mean Square Error (MSE) between the input XX and the reconstructed XX^{\wedge}.

In this work, we proposed the training of an autoencoder to minimize the loss function MSE(X,N)MSE\left(X^{\wedge},N\right), illustrated in Figure 1, where NN represents a face image vector similar to XX and X=X^{\wedge}= decoder (encoder(X))(\operatorname{encoder}(X)). We put forth a method for the automatic selection of similar face image vectors. For each training input face image vector XX, multiple similar face image vectors can be taken into consideration. Thus, each images in the dataset could have different number of similar images. In Section 4, we will conduct a comparative analysis between the results obtained from our proposed training approach and the conventional training method, which involves reconstructing the same training face image vectors.

II-B Selection of similar face image vectors

All the training face image vectors are assessed against one another using the cosine scoring technique. For each face image vector, a group of similar face image vectors is chosen. A simple method for selecting these similar face image vectors involves setting a threshold on the cosine scores between the training face image vectors. Face image vectors with scores surpassing this threshold are designated as similar face image vectors. It’s worth noting that this approach may result in a varying number of similar face image vectors for each training face image vector.

Another approach is to adopt a constant value, denoted as k, and choose k similar face image vectors for each training face image vector. This method ensures a consistent number of similar face images for every training face image vector, resulting in a balanced training approach.

Figure 2 provides a visual representation of the process for selecting similar face image vectors when a constant kk number of neighbors is considered. Let XiX_{i} be a training face image vector, where i=(1,,n)i=(1,\ldots,n) and nn represents the total number of training face image vectors. Initially, we compute cosine scores for all training face image vectors relative to one another. Subsequently, we pick the top kk face image vectors with the highest scores to serve as similar face image vectors for each XiX_{i}. These selected similar face image vectors are denoted as NijN_{ij}, where NijN_{ij} is the jth j^{\text{th }} similar of ith i^{\text{th }} training face image vectors.

Thus, we have a total of n×(k1)n\times(k-1) training data, which automatically generated on the fly which address the issue of having large training data. We run extensive experiment to determine the the values of threshold and kk. Thus, the autoencoder is able to learn information about the intrinsic variability of face image vectors, without necessarily using actual face image labels.

Consequently, we accumulate a total of n×(k1)n\times(k-1) samples for the autoencoder training process. The specific values for the threshold and kk are established through experimental investigation, and a detailed discussion on these parameters will be provided in Section 4. Through this approach, the autoencoder effectively assimilates information regarding the session variability inherent in face image vectors, all without the dependency on actual face image labels.

II-C Autoencoder vector extraction

After the autoencoder is trained with the chosen neighbor face image vectors, we proceed to convert the testing face image vectors into a fresh face image vector representation, utilizing the autoencoder, as illustrated in Figure 1. At this stage, we extract the target face image vectors from the output of the autoencoder. In the experimental trials, these extracted vectors have demonstrated an enhanced discriminatory capability for face image vectors, all achieved without reliance on face image labels. Leveraging these extracted vectors, we carry out the experimental trials using the cosine scoring technique.

Refer to caption
Figure 2: Proposed method of selecting the neighbor face image.

III Experimental Setup and Dataset

III-A Experimental Setup

For the training of our model, we utilized the Keras deep learning library [29]. The architecture of the autoencoder employed in this work comprises three hidden layers. This design maintains symmetry between the encoder and decoder sections. Specifically, both the first and third hidden layers contain an identical number of neurons, as illustrated in Fig. 2. Each of the encoder input and decoder output layers consists of 112 by 112 neurons. In the second layer of both the encoder and decoder, there are 800 neurons. The encoder’s output layer is composed of 300 neurons, determining the size of the resulting embedding vector. Within the DNN, the dimension of the face image embedding layer was fixed at 400, while the classification layer encompassed 1000 neurons.

The training of the autoencoder was conducted over 400 epochs or until the error stops decreasing and showed no further decrease. In all layers of the autoencoder, we applied the Rectified Linear Unit (ReLU) activation function, with the exception of the final layer, which utilized a linear activation function. The learning rate was set to 0.03, incorporating a decay rate of 0.0002. We implemented a logarithmically decaying learning rate, spanning from 10210^{-2} to 10810^{-8}. The batch size was configured at 100.

III-B Datasets

The CelebA dataset, introduced by Liu et al. [9], encompasses a vast collection of over 200,000 images featuring 10,177 celebrities. This dataset incorporates diverse elements such as pose variations and background clutter, providing a comprehensive representation of real-world scenarios. To facilitate model training and evaluation, the dataset is partitioned into distinct sets for training, validation, and testing purposes. In this study, we leveraged the training and test segments of the CelebA dataset as our unlabeled dataset for training the autoencoder. In contrast, supervised training was conducted using the validation portion of the CelebA dataset.

The Labeled Faces in the Wild dataset (LFW) [30] comprises a collection of 13,233 images featuring 5,749 distinct individuals. For testing purposes, this database is randomly and uniformly divided into ten subsets. Within each subset, 300 pairs of matched images (depicting the same person) and 300 pairs of mismatched images (depicting different persons) are randomly selected. In essence, this testing protocol employs a total of 3,000 matched pairs and 3,000 mismatched pairs [30].

IV Experimental Results

IV-A Comparison with different threshold values

TABLE I: The Equal Error Rate (EER) was computed for both the proposed face image vectors and the baseline, assessed across various threshold values when employing cosine scoring.
Approach threshold EER(%)
[1] Baseline No 16.84
[2] Proposed 0.7 15.30
[3] Proposed 0.6 15.04
[4] Proposed 0.5 13.82
[5] Proposed 0.4 11.23
[6] Proposed 0.3 9.61
[7] Proposed 0.2 8.13
[8] Proposed 0.1 8.24
Fusion of [1] and [7] combined 7.12

In Tables I and II, approach refers to the method where face image vectors are extracted using a conventionally trained autoencoder. Table I presents a performance comparison between our proposed method and the baseline, where a threshold is set to select a specific number of neighbor face image vectors. All vectors are then evaluated using the cosine scoring technique. We experimented with different threshold values to fine-tune the system and achieve the best results. The table clearly illustrates that our proposed method outperforms the baseline system. As we decrease the threshold value, the system’s performance improves. The optimal Equal Error Rate (EER) of 8.24% was achieved with a threshold set to 0.2. This represents a 51% relative improvement over the baseline method. Consequently, this bridges the performance gap between cosine and Probabilistic Linear Discriminant Analysis (PLDA) scoring techniques.

By performing a score-level fusion of the baseline and our proposed method, using a threshold of 0.2, we achieved a further enhancement in performance, as evidenced by a reduced Equal Error Rate (EER) of 7.12%. The optimal weights for the fusion were determined empirically, resulting in values of 0.55 and 0.45 for the baseline and our proposed method, respectively. This fusion strategy effectively leverages the strengths of both approaches to yield superior results.

IV-B Comparison with different k values

TABLE II: The Equal Error Rate (EER) was computed for both the proposed face image vector and the baseline method across different values of k all evaluated using the cosine scoring technique.
Approach k EER(%)
[1] Baseline No 18.54
[2] Proposed 1 15.21
[3] Proposed 2 13.13
[4] Proposed 3 12.71
[5] Proposed 4 10.12
[6] Proposed 10 8.40
[7] Proposed 11 8.03
[8] Proposed 12 8.94
Fusion of [1] and [11] combined 7.01

In Table II, we present a performance comparison between our proposed method and the baseline, utilizing a constant value of k for the selection of similar face image vectors. We conducted experiments with various values of k to optimize the system for achieving the best results. The table illustrates that employing a fixed value of k has demonstrated improvement compared to both the threshold-based approach and the baseline. Beginning with k set to 1, an Equal Error Rate (EER) of 15.21% was attained. This signifies a 17% relative improvement over the baseline system by considering only the nearest neighbor for each image vector. As we increment the value of k, the system’s performance continues to improve. The most favorable EER of 8.03% was achieved when k was set to 11, leading to a remarkable 56% relative improvement over the baseline method. This achievement effectively bridges the performance gap between cosine and PLDA scoring techniques. Moreover, this approach enables a balanced training regimen, ultimately enhancing the system’s performance. It is worth noting that a further increase in the value of k may incorporate very distant neighbors, potentially degrading the system’s performance.

By combining the scores from the baseline and our proposed method, both utilizing k set to 11, we achieved a further improvement in performance, specifically in terms of the Equal Error Rate (EER). This fusion approach yielded an EER of 7.01%, which is in close proximity to the results obtained using Probabilistic Linear Discriminant Analysis (PLDA) with actual face image labels. The optimal weights for the fusion were fine-tuned through experimental iteration and were ultimately set to 0.49 and 0.51 for the baseline and our proposed method, respectively. This fusion strategy effectively leverages the strengths of both approaches to yield highly competitive results.

IV-C Comparison with score level fusion

TABLE III: The Equal Error Rate (EER) was computed for the score-level fusion between Probabilistic Linear Discriminant Analysis (PLDA) and cosine scoring, using both the proposed face image vectors with k set to 15, and the baseline method.
Approach Scoring EER(%)
[1] Baseline PLDA 6.71
[2] Proposed (k = 11) Cosine 8.03
Fusion of [1] and [2] combined 6.23

As depicted in Table III, conducting a score-level fusion between Probabilistic Linear Discriminant Analysis (PLDA) and cosine scoring, utilizing the proposed method with k set to 11 (yielding the best results), leads to an Equal Error Rate (EER) of 6.23%. This showcases an impressive relative improvement of nearly 7% over the PLDA approach alone. The optimal fusion weights were determined to be 0.04 for the baseline and 0.96 for the proposed method, respectively. This fusion strategy effectively capitalizes on the strengths of both approaches, resulting in a highly competitive performance.

IV-D Comparison with the state of the art on LFW dataset

TABLE IV: Comparison of the proposed method result with some of state-of-the-art methods on LFW testing dataset.
Model Training size Labeled/ Unlabeled Testing size Accuracy(%)
UniformFace 6.1M Labeled 6K 99.80 [31]
ArcFace 5.8M Labeled 6K 99.82 [32]
GroupFace 5.8M labeled 6K 99.85 [33]
CosFace 5M Labeled 6K 99.73 [34]
Marginal Loss 4M Labeled 6K 99.48 [35]
CurricularFace 3.8M Labeled 6K 99.80 [36]
RegularFace 3.1M Labeled 6K 99.61 [37]
AFRN 3.1M Labeled 6K 99.85 [38]
Stream Loss 1.5M Labeled 6K 98.97 [39]
MDCNN 1M Labeled 6K 99.38 [40]
ULNet 1M Labeled 6K 99.70 [41]
Ben Face 0.5M Labeled 6K 99.20 [42]
F2C 5.8M Labeled 6K 99.83 [43]
PCCycleGAN 0.5M Unlabeled 6K 99.52 [44]
CAPG GAN 1M Unlabeled 6K 99.37 [45]
UFace 200K Unlabeled 6K 99.40 [46]
Proposed 200K Unlabeled 6K 99.53

Despite the slightly higher accuracy of methods like ArcFace, GroupFace, Marginal Loss, and CosFace, it’s noteworthy that our proposed method achieves competitive results with a significantly smaller training dataset (around 200K images). In contrast, many state-of-the-art methods rely on much larger training sets, often in the millions of images. This demonstrates the efficiency and effectiveness of our approach, particularly in scenarios where acquiring a massive dataset may not be feasible or practical.

The experimental result of the proposed method become particularly notable when compared to state-of-the-art approaches. For instance, ArcFace, which attains 99.82% accuracy, utilized a substantially larger dataset of 5.8 million labeled images. In contrast, our method achieved comparable performance with significantly fewer training images (around 200K). This underscores the efficiency and effectiveness of our approach, especially in scenarios where obtaining a massive dataset is challenging.

V Conclusion

The proposed work introduces a novel unsupervised deep learning method for face verification. The training process involves selecting k images from unlabeled data that are most similar to a specified face image, and using them for training. Unlike standard deep neural network training, our proposed method calculates the loss function for each input image k times. The primary goal is to bridge the performance gap between cosine and PLDA scoring techniques when labeled training data is unavailable. This is achieved by training an autoencoder to reconstruct a set of neighbor face image vectors, instead of reconstructing the same training face image vectors. These neighbor face image vectors are selected based on their cosine scores in relation to the training face image vector. After training, face image vectors are extracted for test face images as the output of the autoencoder. In experimental trials, face image vectors are scored using cosine scoring. The evaluation conducted on the LFW dataset demonstrates that our proposed method achieves a relative improvement of 56% in terms of Equal Error Rate (EER) over the baseline system. This successfully narrows the performance gap between cosine and PLDA scoring systems.

References

  • [1] Jain, A.; Li, S. Handbook of Face Recognition; Springer: Berlin/Heidelberg, Germany,
  • [2] Solomon, E. Face Anti-Spoofing and Deep Learning Based Unsupervised Image Recognition Systems. (2023)
  • [3] Solomon, E. Face Anti-Spoofing and Deep Learning Based Unsupervised Image Recognition Systems. VCU Theses And Dissertations. (2023), https://scholarscompass.vcu.edu/etd/7482
  • [4] Solomon, E. & Cios, K. FASS: Face Anti-Spoofing System Using Image Quality Features and Deep Learning. Electronics. 12, 2199 (2023)
  • [5] Solomon, E. & Cios, K. HDLHC: Hybrid Face Anti-Spoofing Method Concatenating Deep Learning and Hand-Crafted Features. 2023 IEEE 6th International Conference On Electronic Information And Communication Technology (ICEICT). pp. 470-474 (2023)
  • [6] Solomon, E., Woubie, A. & Emiru, E. Deep Learning Based Face Recognition Method using Siamese Network. ArXiv Preprint ArXiv:2312.14001. (2023)
  • [7] Solomon, E., Woubie, A. & Emiru, E. Autoencoder Based Face Verification System. ArXiv Preprint ArXiv:2312.14301. (2023)
  • [8] Woubie, A., Solomon, E. & Emiru, E. Image Clustering using Restricted Boltzman Machine. ArXiv Preprint ArXiv:2312.13845. (2023)
  • [9] Liu, Z., Luo, P., Wang, X. & Tang, X. Deep learning face attributes in the wild. Proceedings Of The IEEE International Conference On Computer Vision. pp. 3730-3738 (2015)
  • [10] Deng, L., Yu, D. & Others Deep learning: methods and applications. Foundations And Trends® In Signal Processing. 7, 197-387 (2014)
  • [11] Snyder, D., Garcia-Romero, D., Sell, G., Povey, D. & Khudanpur, S. X-vectors: Robust dnn embeddings for speaker recognition. 2018 IEEE International Conference On Acoustics, Speech And Signal Processing (ICASSP). pp. 5329-5333 (2018)
  • [12] Abraham Woubie and Tom Bäckström: Federated learning for privacy-preserving speaker recognition, In IEEE Access, 2021; pp. 149477–149485.
  • [13] Abraham Woubie and Tom Bäckström: Federated Learning for Privacy Preserving On-Device Speaker Recognition, In ISCA Symposium on Security and Privacy in Speech Communication, 2021; pp. 1–5.
  • [14] Abraham Woubie, Pablo Zarazaga and Tom Bäckström: The Use of Audio Fingerprints for Authentication of Speakers on Speech Operated Interfaces, In ISCA Symposium on Security and Privacy in Speech Communication, 2021; pp. 6–9.
  • [15] Abraham Woubie and Tom Bäckström: Voice-quality Features for Deep Neural Network Based Speaker Verification Systems, In 29th European Signal Processing Conference (EUSIPCO), 2021; pp. 176–180.
  • [16] Chekole, E., Chattopadhyay, S., Ochoa, M., Guo, H. & Cheramangalath, U. CIMA: Compiler-Enforced Resilience Against Memory Safety Attacks in Cyber-Physical Systems. Computers & Security. 94 pp. 101832 (2020), https://www.sciencedirect.com/science/article/pii/S0167404820301061
  • [17] Chekole, E., Ochoa, M. & Chattopadhyay, S. SCOPE: Secure Compiling of PLCs in Cyber-Physical Systems. International Journal Of Critical Infrastructure Protection. 33 pp. 100431 (2021), https://www.sciencedirect.com/science/article/pii/S1874548221000238
  • [18] Chekole, E. & Huaqun, G. ICS-SEA: Formally Modeling the Conflicting Design Constraints in ICS. Proceedings Of The Fifth Annual Industrial Control System Security (ICSS) Workshop. pp. 60-69 (2019), https://doi.org/10.1145/3372318.3372325
  • [19] Chekole, E., Castellanos, J., Ochoa, M. & Yau, D. Enforcing Memory Safety in Cyber-Physical Systems. Computer Security. pp. 127-144 (2018)
  • [20] Chekole, E., Chattopadhyay, S., Ochoa, M. & Huaqun, G. Enforcing Full-Stack Memory-Safety in Cyber-Physical Systems. Engineering Secure Software And Systems. pp. 9-26 (2018)
  • [21] Chekole, E., Thulasiraman, R. & Zhou, J. EARIC: Exploiting ADC Registers in IoT and Control Systems. Applied Cryptography And Network Security Workshops. pp. 245-265 (2023)
  • [22] Wong, A., Chekole, E., Ochoa, M. & Zhou, J. On the Security of Containers: Threat Modeling, Attack Analysis, and Mitigation Strategies. Computers & Security. 128 pp. 103140 (2023), https://www.sciencedirect.com/science/article/pii/S0167404823000500
  • [23] Liu, P., Han, S., Meng, Z. & Tong, Y. Facial expression recognition via a boosted deep belief network. Proceedings Of The IEEE Conference On Computer Vision And Pattern Recognition. pp. 1805-1812 (2014)
  • [24] Kurup, A., Ajith, M. & Ramón, M. Semi-supervised facial expression recognition using reduced spatial features and deep belief networks. Neurocomputing. 367 pp. 188-197 (2019)
  • [25] Teh, Y. & Hinton, G. Rate-coded restricted Boltzmann machines for face recognition. Advances In Neural Information Processing Systems. 13 (2000)
  • [26] Alphonse, A., Shankar, K., Jeyasheela Rakkini, M., Ananthakrishnan, S., Athisayamani, S., Robert Singh, A. & Gobi, R. A multi-scale and rotation-invariant phase pattern (MRIPP) and a stack of restricted Boltzmann machine (RBM) with preprocessing for facial expression classification. Journal Of Ambient Intelligence And Humanized Computing. 12 pp. 3447-3463 (2021)
  • [27] Nguyen, H. & Bai, L. Cosine similarity metric learning for face verification. Asian Conference On Computer Vision. pp. 709-720 (2010)
  • [28] Muslihah, I. & Muqorobin, M. Texture characteristic of local binary pattern on face recognition with probabilistic linear discriminant analysis. International Journal Of Computer And Information System (IJCIS). 1, 22-26 (2020)
  • [29] Chollet, F. & Others Keras. https:// keras. io. (2015)
  • [30] Huang, G., Mattar, M., Berg, T. & Learned-Miller, E. Labeled faces in the wild: A database forstudying face recognition in unconstrained environments. (2008)
  • [31] Duan, Y., Lu, J. & Zhou, J. Uniformface: Learning deep equidistributed representation for face recognition. Proceedings Of The IEEE/CVF Conference On Computer Vision And Pattern Recognition. pp. 3415-3424 (2019)
  • [32] Deng, J., Guo, J., Xue, N. & Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. Proceedings Of The IEEE/CVF Conference On Computer Vision And Pattern Recognition. pp. 4690-4699 (2019)
  • [33] Kim, Y., Park, W., Roh, M. & Shin, J. Groupface: Learning latent groups and constructing group-based representations for face recognition. Proceedings Of The IEEE/CVF Conference On Computer Vision And Pattern Recognition. pp. 5621-5630 (2020)
  • [34] Wang, H., Wang, Y., Zhou, Z., Ji, X., Gong, D., Zhou, J., Li, Z. & Liu, W. Cosface: Large margin cosine loss for deep face recognition. Proceedings Of The IEEE Conference On Computer Vision And Pattern Recognition. pp. 5265-5274 (2018)
  • [35] Deng, J., Zhou, Y. & Zafeiriou, S. Marginal loss for deep face recognition. Proceedings Of The IEEE Conference On Computer Vision And Pattern Recognition Workshops. pp. 60-68 (2017)
  • [36] Huang, Y., Wang, Y., Tai, Y., Liu, X., Shen, P., Li, S., Li, J. & Huang, F. Curricularface: adaptive curriculum learning loss for deep face recognition. Proceedings Of The IEEE/CVF Conference On Computer Vision And Pattern Recognition. pp. 5901-5910 (2020)
  • [37] Zhao, K., Xu, J. & Cheng, M. Regularface: Deep face recognition via exclusive regularization. Proceedings Of The IEEE/CVF Conference On Computer Vision And Pattern Recognition. pp. 1136-1144 (2019)
  • [38] Kang, B., Kim, Y., Jun, B. & Kim, D. Attentional feature-pair relation networks for accurate face recognition. Proceedings Of The IEEE/CVF International Conference On Computer Vision. pp. 5472-5481 (2019)
  • [39] Rashedi, E., Barati, E., Nokleby, M. & Chen, X. “Stream loss”: ConvNet learning for face verification using unlabeled videos in the wild. Neurocomputing. 329 pp. 311-319 (2019)
  • [40] Huang, X.; Zeng, X.; Wu, Q.; Lu, Y.; Huang, X.; Zheng, H. Face Verification Based on Deep Learning for Person Tracking in Hazardous Goods Factories. Processes 2022, 10, 380.
  • [41] Boragule, A.; Akram, H.; Kim, J.; Jeon, M. Learning to Resolve Uncertainties for Large-Scale Face Recognition. Pattern Recognit. Lett. 2022, 160, 58–65.
  • [42] Ben Fredj, H.; Bouguezzi, S.; Souani, C. Face recognition in unconstrained environment with CNN. Vis. Comput. 2021, 37, 217–226.
  • [43] Wang, K.; Wang, S.; Zhang, P.; Zhou, Z.; Zhu, Z.; Wang, X.; Peng, X.; Sun, B.; Li, H.; You, Y. An efficient training approach for very large scale face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, Louisiana, USA, 21-24 June 2022; pp. 4083–4092.
  • [44] Liu, Y.; Chen, J. Unsupervised face frontalization for pose-invariant face recognition. Image Vis. Comput. 2021, 106, 104093.
  • [45] Hu, Y.; Wu, X.; Yu, B.; He, R.; Sun, Z. Pose-guided photorealistic face rotation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18-23 June 2018; pp. 8398–8406.
  • [46] Solomon, E., Woubie, A. & Cios, K. UFace: An Unsupervised Deep Learning Face Verification System. Electronics. 11, 3909 (2022)