Locality-aware Channel-wise Dropout for Occluded Face Recognition
Abstract
Face recognition remains a challenging task in unconstrained scenarios, especially when faces are partially occluded. To improve the robustness against occlusion, augmenting the training images with artificial occlusions has been proved as a useful approach. However, these artificial occlusions are commonly generated by adding a black rectangle or several object templates including sunglasses, scarfs and phones, which cannot well simulate the realistic occlusions. In this paper, based on the argument that the occlusion essentially damages a group of neurons, we propose a novel and elegant occlusion-simulation method via dropping the activations of a group of neurons in some elaborately selected channel. Specifically, we first employ a spatial regularization to encourage each feature channel to respond to local and different face regions. In this way, the activations affected by an occlusion in a local region are more likely to be located in a single feature channel. Then, the locality-aware channel-wise dropout (LCD) is designed to simulate the occlusion by dropping out the entire feature channel. Furthermore, by randomly dropping out several feature channels, our method can well simulate the occlusion of larger area. The proposed LCD can encourage its succeeding layers to minimize the intra-class feature variance caused by occlusions, thus leading to improved robustness against occlusion. In addition, we design an auxiliary spatial attention module by learning a channel-wise attention vector to reweight the feature channels, which improves the contributions of non-occluded regions. Extensive experiments on various benchmarks show that the proposed method outperforms state-of-the-art methods with a remarkable improvement.
Index Terms:
Occluded face recognition, locality-aware channel-wise dropout, spatial attention module.I Introduction
With the huge success of deep learning, a remarkable improvement has been achieved for face recognition under controlled settings (i.e., occlusion-free images, near-frontal poses, neutral expressions, normal illuminations, etc.). However, in realistic unconstrained scenarios, face recognition remains a challenging task due to various factors including very large poses, very low resolution and occlusions. Among these factors, the occlusion is an intractable problem which leads to a severe degeneration in recognition accuracy.
The occlusions always bring about two primary issues, i.e., the missing of facial information and the noise from occlusion. To improve the robustness against occlusion, many efforts [1, 2, 3, 4, 5, 6] have been made to recover the occluded faces. The work in [2] uses a multi-scale spatial long short-term memory (LSTM) encoder to encodes occluded face patches, and then another LSTM is employed to reconstruct the occlusion-free face image. Based on the Generative Adversarial Network (GAN) [7], another work [5] proposes a face completion model to generate visually plausible contents for the occluded face regions. Although a huge progress has been made, the performance of occlusion removal is still far from satisfactory. The main reason is that these methods are commonly trained with artificial occluded images. For instance, the images are manually generated by randomly putting a black rectangle or several object templates including sunglasses, scarfs, phones, and cups on them, which differ significantly from real-life occlusions. Therefore these methods always suffer from poor generalizations under realistic scenarios. Besides, how to well recover the occluded face regions as well as preserve the identity information is another challenge for these methods.
Another type of methods focus on suppressing the noise caused by occlusions. They attempt to discard the corrupted feature elements which are extracted from the occluded regions [8, 9, 10, 11, 12, 13]. PDSN [12] builds a mask dictionary in advance by comparing the features extracted from normal faces and those from occluded faces. During the recognition process, the occluded facial regions are detected by a segmentation network and then the noise is removed by discarding the corrupted feature elements retrieved from the mask dictionary. To some extent, the occlusion discarding method relieves the influence of occlusions. However it may be non-trivial to precisely detect the real-world occlusions which usually have various shapes and textures since the occlusion detection modules are also trained with artificial occlusions. Even if the corrupted feature elements have been located perfectly, directly zeroing out them will incorporate a peculiar pattern into the final feature representation. Moreover, due to the uncertain size of the occluded region, the final feature will have an unfixed number of valid elements. Thus, the traditional metrics designed for the fixed-length vectors may fail for evaluations.

One simple way of tackling occluded face recognition problem is to leverage a massive number of occluded faces in the real world to directly train deep neural networks, which are forced to learn occlusion-robust face features. However, it is hard to collect such a training set. As an alternative, augmenting the training images with artificially synthesized occlusions has been studied and significant improvements have been witnessed in [14, 15, 16, 17]. However, the image-level occlusion simulation is still not an elegant solution since the artificial occluded images are also generated by using limited hand-craft occlusion templates, which cannot fully represent the arbitrary patterns in realistic occlusions.
The common issue of all the aforementioned methods is that the artificially synthesized occlusion cannot precisely represent the real-world occlusions, leading to poor generalizations under realistic scenarios. In this paper, we propose a novel and elegant method which can better simulate realistic occlusions. Here, we hold the opinion that the occlusion essentially damages a group of neurons. To synthesize different occlusions, a natural approach is to dropout the activations of various neurons. However, conventional dropout operation cannot simulate the real-life occlusion. The reason is that real-life occlusion affects a contiguous region in an activation map, while the conventional dropout operation discards discrete activations. To better simulate the feature damaged by occlusions, we propose the revised dropout method, namely the locality-aware channel-wise dropout (LCD) to drop a group of activations which are affected by the same facial occlusion. For conventional neural networks, partial occlusion usually affects activations across multiple feature channels, which makes it difficult to simulate occlusions with a few channels. Inspired by the [18], we encourage each feature channel to only respond to a local and different face region via the spatial regularization. As shown in Fig. 1, the heat maps of 3 different feature channels are visualized and it can been seen that these channels respond to various face regions. Therefore, the activations affected by a partial occlusion are more likely to be located in a single feature channel. Then, LCD can simulate the occlusion at the local region by dropping out the entire feature channel. It can also simulate the occlusion of larger area by randomly dropping out more feature channels.
Setting our LCD at a middle depth in the neural network can encourage its succeeding layers to minimize the intra-class feature variance caused by occlusions, leading to improved robustness against occlusion. Furthermore, we design an auxiliary spatial attention module which learns a channel-wise attention vector to reweight the channels during the feature extraction process. After jointly trained with our LCD, the deep network is optimized to focus more on the channels which are related to the non-occluded regions and suppress others which are affected by the occluded regions.
Compared with previous works, our method has three major advantages: 1) our method does not require artificially synthesized occlusions. Instead, it gracefully simulates the realistic occlusions in intermediate features. 2) Different from previous works which use additional module to detect or recover the occluded region, our method imposes minor increase in model complexity during the inference phase. 3) Our method is a more practical approach which can be seamlessly integrated with any existing face recognition method for improving robustness to occlusions.
The main contributions of this paper are summarized as follows:
-
•
We propose a novel method to better simulate realistic occlusions by dropping a group of activations in intermediate features. It significantly improves the robustness to occlusions by encouraging the neural network to emphasize on learning discriminative features from the non-occluded face regions.
-
•
An auxiliary spatial attention module is designed to improve the contributions of non-occluded regions by adaptively reweighting the feature channels.
-
•
Our method significantly outperforms the state-of-the-art methods on IJB-C, LFW and MegaFace benchmarks, especially on the IJB-C dataset with large-scale real-occluded face images.
The rest of the paper is organized as follows. Section II gives a brief overview of the related works. Section III introduces the detailed formulation of the proposed method, followed by a discussion with other state-of-the-art methods in Section IV. Section V presents the ablation study and the experimental results on three databases. Finally, the conclusion is summarized in Section VI.
II Related Works
The existing occlusion robust face recognition methods can be broadly grouped into the following three categories: occluded face completion methods, occlusion-aware discarding methods and occlusion-robust feature extraction methods. In this section, we provide a brief overview of the recent works which are most relevant to this paper.
II-A Occluded Face Completion Methods
The occluded face completion methods are pixel-level approaches which aim to recover the occluded face regions. Considering the low-rank property of non-occluded images, early works attempt to solve the problem by using robust principal component analysis (PCA) to reconstruct the corrupted low-rank face images [19, 20, 21, 22, 23]. The work in [21] proposes the robust principal component analysis (RPCA) which improves the performance of removing shadows from face images. In [22], an important extension of RPCA, namely the low-rank representation (LRR) is presented to extend the recovery of corrupted images from a single subspace to multiple subspaces. Both RPCA and LRR assume that the occluded pixels are sparse, while real-world face images usually contain dense occlusions, which make the matrix non-sparse. To this end, [23] presents the double nuclear norm-based matrix decomposition to remove the dense occlusions.
Recently, more works resort to the deep learning for improving the occluded face completion [1, 2, 3, 4, 5, 6, 24]. In [1], the difference between the activation values of two stacked sparse denoising auto-encoders (SSDAs) are used to indicate occluded and un-occluded face regions. Then, the final occlusion-free image is reconstructed by transferring the encoding activations of the un-occluded region to the occluded region. In [3], the context encoder combines the auto-encoder architecture with context information of the occluded part to produce visually pleasing images. To further improve the context encoder, [4] introduces both global and local discriminators. Specifically, the global discriminator pursues the global consistency of the overall image and the local discriminator looks at a small area centered at the reconstructed region to judge the quality in details. To ensure the new generated contents more photo-realistic, a semantic parsing loss is developed in [5]. To refine local face textures, a 3D morphable model (3DMM) is utilized in [6] to further assist the learning of the local discriminator. In the unsupervised face normalization method (FNM) [24], multiple local discriminators are integrated into a novel unsupervised framework. It generates impressive high quality faces which have dispelled various face variations including occlusions.
Overall, the above-mentioned face completion methods have shown promising results of transforming occluded faces to un-occluded ones. However, two major issues still remain. Firstly, except for a few methods (e.g., the FNM [24]), most previous methods require pair-wise training data (i.e., one occluded face and one un-occluded face of the same person). Unfortunately, such training sets containing a mass of pair-wise faces with natural occlusion are extremely rare. An alternative and commonly used approach is using synthetically occluded faces. However, these synthetically occluded faces, e.g., using manually designed occlusion templates or a black/white rectangle can not fully represent natural occluded faces. Secondly, the faces generated by GAN-based methods are usually visually pleasing. However, how to remove the occlusion while preserving the identity information still remains a challenging problem.
II-B Occlusion-Aware Discarding Methods
These approaches aims to remove the noise caused by occlusions with two pipelines, which either discards the occluded pixels before the face feature extraction, or discards the corrupted feature elements during the feature extraction.
Following the former pipeline, some works detect the occlusions first and then extract a feature representation from the non-occluded regions only. The early works [25, 8, 9, 10] usually employ a nearest neighbor classifier (NNC) or a support vector machine (SVM) to classify the occluded face regions. Since these methods are designed based on the traditional feature descriptors, it is non-trivial for them to obtain discriminative ability for face recognition in complex scenarios. Recently, the LPD [13] designs a neural network to locate the latent facial parts which are less affected by a specific occlusion (i.e., the respirator), and then extracts discriminative features from the selected latent part.
Following the spirit of the latter pipeline, the mask leaning methods [11, 12] locate and discard the corrupted feature elements rather than the occluded pixels. In [11], the MaskNet adaptively learns feature masks for occluded face images and automatically assigns lower weight to the hidden units activated by the occluded face regions. The PDSN [12] establishes a mask dictionary to represent the correspondence between the occluded facial block and the corrupted feature elements. During the testing phase, a segmentation network is employed to detect the occluded facial blocks, and then the corrupted feature elements are set to zero by retrieving the relevant dictionary items.
Although above-mentioned occlusion-aware discarding methods can alleviate the occlusion issue, completely discarding the occluded regions still takes the risk of reducing the system reliability. Firstly, precise and fine-grain occlusion detection is an essential prerequisite for these methods. However, such occlusion detection is non-trivial to obtain and a coarse detection will increase the risk of losing discriminative information or inducing unreliable information. Furthermore, even if the occluded face regions or the corrupted feature elements have been located perfectly, directly zeroing out the corrupted elements still incorporates a peculiar pattern into the final feature, which may harm the recognition accuracy. Moreover, due to the arbitrary occlusion, the feature vectors for occluded faces have unfixed number of valid elements. Thus, the conventional metrics for fixed-length vectors are not fully applicable. Although works on the instance-to-class distance [26] and the reconstruction-based similarity measurement [27] are proposed to tackle the above problem, the commonly used window sliding makes them relatively time-consuming. On the other side, directly ignoring occluded elements will potentially break the global cues of face images such as chin contours, which is also harmful to face recognition system.
II-C Occlusion-Robust Feature Extraction Methods
Directly learning occlusion-robust feature is the most straightforward and effective way to handle occlusion face recognition. For this purpose, many efforts are devoted to seeking a feature space that is less affected by occlusion and meanwhile preserves the discriminative capability of distinguishing identities. Many works seek such feature space via the sparse representation classification (SRC) [28], in which the occluded image is represented by a linear combination of training samples plus a sparse constraint term accounting for occlusions. The LR-LUM [29] combines both the robust sparsity constraint and the low-rank constraint which outperforms previous methods in handing structured occlusion such as sunglasses and scarfs. Other works [30, 31, 32] extend the sparse representation by combining more discriminative feature descriptors. In [32], the JCR-ACF proposes a joint and collaborative representation with local adaptive convolution feature, which can improve the recognition accuracy by employing information from different local face regions. Although these SRC-based methods have made considerable progresses, they do not generalize well in practical scenarios since they requires that the test samples have identical identities with a pre-defined close-set.
Owing to leveraging massive training sets, recent deep learning methods reveal significant superiority on face recognition. Enlarging the training datasets with sufficient occluded faces may be an effective way to improve the occlusion robustness of face embedding. Unfortunately, such training sets of a mass of identities are extremely rare. As an alternative solution, synthetic image augmentation method has been studied by previous works [14, 15, 16, 17]. The work in [14] enriches the training set by synthesizing occluded faces with various pre-defined hairstyles and glasses templates. Although an accuracy gain has been witnessed, the diversity of the manually designed occlusion templates needs to be further improved. More recently, BFL [17] proposes an enhanced augmentation schema to randomly generate multi-scale spatially occluded samples and then modifies the loss to balance the impact of normal and occluded samples for training. To some extent, these augmentation methods improve the robustness of neural network, but the discrepancy between the synthetic and real occluded faces still limits the further improvements of the robustness.
Another line of researches [33, 34, 35] focus on the attention mechanism for robust feature extraction. The state-of-the-art method named InterpretFR [35] employs a Siamese network to compare the feature elements from a normal face with and without synthetic occlusion. Then it encourages the neural network to identify the input face solely based on the feature elements which are less sensitive to occlusions. However, it still need synthesizing occluded faces by using artificial occlusion templates and fails to generalize well on unseen occlusions. In contrast, our method can realistically simulate arbitrary occlusions via dropping out a random group of filter responses, leading to an improved performance under real world scenarios.
III METHODOLOGY
III-A Overview
The proposed method attempts to learn occlusion robust face features by simulating occlusions during the training process. Different from previous methods which augment face images with synthesized occlusions, we propose to directly simulate the influence of arbitrary occlusions on intermediate features. As the occlusion essentially damages a group of neurons, we propose a revised dropout method, namely the locality-aware channel-wise dropout (LCD) to simulate occlusions by dropping a group of feature channels. Considering that, for the conventional neural networks, the occlusion usually affects the activations from most channels which is even impossible to simulate the occlusions by dropping several channels. Therefore, we first employ a spatial regularization to encourage each feature channel to respond to different face regions. In this way, the activations affected by the partial occlusion are more likely to be located in a single feature channel. Then, our LCD can simulate the occlusion by dropping out the entire feature channel, in which sense we name this method as locality-aware channel-wise dropout. Moreover, to improve the contributions of non-occluded regions for learning occlusion robust features, we design an auxiliary spatial attention module to reweight the feature channels. The whole framework shown in Fig. 1 is end-to-end trainable and can be easily applied on any existing convolution neural network.
III-B Locality-aware Channel-wise Dropout
III-B1 Spatial regularization of feature channels
We employ the spatial regularization [18] to encourage each feature channel to respond to local and different face regions. The spatial regularization consists of two loss functions. The first one is a filter orthogonal loss:
(1) |
where denotes the -th column of convolution filter . The Eq. 1 encourages the orthogonality of filters, which make the different filters more likely to respond to various face regions. Besides, the other loss function is employed to further enhance such characteristic by directly penalizing the correlations between the filters’ responses:
(2) |
where denotes the response of -th filter (i.e., the -th channel of features).
III-B2 Simulating occlusions via Channel-wise Dropout
Given an input feature map , we first generate an all-one mask matrix with the same size as , where , , denotes the channel number, the height, the width of the feature map, respectively. Second, we randomly sample distinct channel indexes from the channels. Then, the mask values for these channels are set to zero:
(3) |
Finally, the output of the LCD is obtained by the product of the mask matrix and the input:
(4) |
where denotes Hadamard product. As shown in Fig. 1, all the feature elements within the feature maps will be dropped out to zero. At the training stage, our LCD performs as an effective occlusion simulation and encourages the network to identify the input face solely based on the remaining feature, make it more robust to occlusions. It should be mentioned that similar to the conventional dropout, the LCD is not employed during the inference process.
In our method, is a crucial parameter which controls how many feature channels will be dropped out. In other words, it determines how many face regions will be discarded during the training process. A larger simulates a more severe occlusion where more regions in the face are occluded. To simulate the complex pattern of realistic occlusions, a dynamic for each training samples is required. With this in mind, the is designed as a stochastic variable satisfying uniform distribution within . A larger (e.g., ) is recommended to improve the robustness of severer occlusions.
Since the activations of convolutions layers are commonly normalized by batch-normalization (BN) layers, how to make the LCD compatible with the BN layers is worth exploring. In a conventional BN layer, the feature elements sharing the same channel index will be normalized together. However, this process will have some problems when applying the LCD before the BN layer. Specifically, for a mini-batch with samples, let denote its -th element in the feature tensor. The conventional BN layer computes the mean for -th channel:
(5) |
However, since features of a sub-set channels are set to zero, the number of valid samples for the -th channel is no longer equal to . To resolve this problem, the calculation of must be modified to:
(6) |
where denotes the number of training samples which have zero values in the -th channel. Besides, the calculation of the channel variance also requires similar modification. To be free from these modifications, always setting the LCD after the conventional BN layer is an alternative solution. Moreover, this simple setting is even more favorable to obtain stable BN parameters as all the training sample are involved in the computation of mean and variance.
III-C Spatial Attention Module

In an occluded face image, the feature channels responded to the occluded regions will not contain discriminative information for face recognition. Besides, they potentially contains the noise from the occluded regions. To be free from this, we design an attention module to reweight the feature channels, making the network focus more on the feature channels which are related to the non-occluded regions. It is worth noting that, by utilizing the aforementioned spatial regularization, each feature channel is encouraged to respond to local and different face regions. In this sense, reweighting these feature channels actually performs as a spatial-wise attention approach, which is named as the spatial attention module (SAM).
For the input feature map , our goal is to learn an attention vector which controls the weight of each feature channel. As shown in Fig. 1, a light-weight module is designed to learn the attention vector . Specifically, we first employ a convolution layer and a fully-connected layer to obtain a global view of all input feature maps. Then, another fully-connected layer is utilized to extract the channel-wise attention vector. It is worth noting that previous attention block in SENet [36] utilizes a global average pooling as a global descriptor, while in our method the global pooling of dropped zero value feature maps will suffer from local minimum. To this end, the convolution layer in our proposed module is a necessary and effective information aggregation strategy to achieve attentions.
The refined feature map is obtained by channel-wise multiplication between each feature map and its corresponding attention scalar :
(7) |
Formally, the Algorithm 1 summarizes the training process using the locality-aware channel-wise dropout and the spatial attention module.
Compute the filter orthogonal loss ;
Randomly sample channels from the ;
Backward propagation;
IV Discussion
The proposed LCD method can be categorized as a form of structured dropout. Conventionally, other structured dropout methods are designed for the purposes of alleviating the over-fitting issue and improving the generalization ability. In contrast, our LCD is designed for simulating facial occlusions and achieving occlusion robustness. In this section, we discuss the comparisons of our LCD with two representative structured dropout methods, i.e., the DropBlock [37] and the weighted channel dropout (WCD) [38]. Furthermore, we also provide a comparison with the occluded face recognition method InterpretFR [35] which is most relevant to our method.
IV-A Differences from DropBlock [37]
Both our method and DropBlock drop some activations in the intermediate feature layers, but they differ in two aspects. 1) The Dropblock drops partial regions within each feature channel for enhancing feature learning while our method employs spatial regularization to enforce each feature channel to respond to local face regions and further drop several feature channels to simulate partial occlusions. 2) Our method further proposes spatial attention module to improve the contributions of undamaged neurons to further promote the robustness to occlusion. Experimental results on IJB-C datasets show our method significantly outperforms Dropblock for occluded face recognition.
IV-B Differences from the Weighted Channel Dropout [38]
Both our method and the weighted channel dropout (WCD) drop channels but for different purposes and in different ways. Firstly, in terms of designing purpose, the WCD is for alleviating the over-fitting issue when fine-tuning neural network on small datasets, while our method aims to improve the robustness to partial occlusion for face recognition. Secondly, in terms of what channels to drop, the WCD drops out the feature channels which have relatively lower activation magnitudes, while in our method each feature channel is enforced to respond to features corresponding to some local facial region and thus can be dropped to simulate the feature corruption due to the occlusion of that region. Comparison experiments on IJB-C datasets show the superiority of our method in handling occluded face recognition problem.
IV-C Differences from InterpretFR [35]
Both our method and InterpretFR resort to occlusion robust feature learning for tackling face recognition under occlusions. However, we have two major differences. 1) InterpretFR synthesizes occlusions by putting black boxes on faces, which is unrealistic and the performance may severely degenerate under real world scenarios. Instead of simply utilizing fixed templates to synthesize occlusion, we propose the locality-aware channel-wise dropout to simulate more realistic occlusions, leading to performance improvement for occluded face recognition. 2) During training, InterpretFR masks out the elements of the final face representation sensitive to the occlusion, which may be sub-optimal as it only leverage remaining features for recognition without considering information recovering. Differently, we perform channel-wise dropout on the stage 3 of ResNet and design spatial attention module in the stage 4 to implicitly recover the absent information as verified by experiment in Section V-C, which is more favorable to tackle face recognition under occlusions.
V Experiments
V-A Datasets
The training images are collected from the MS-Celeb-1M dataset [39]. Since this dataset contains many labeling noise, we manually clean it and finally collect 3.7 Million images from 50K identities. The revised dataset is used as our training set for the experiments. We extensively test our method on three popular benchmarks, including IJB-C [40], MegaFace [41] and LFW [42].
IJB-C. As an extension of previous IJB-A [43] and IJB-B [44], the IJB-C [40] is a large-scale dataset which contains 117, 542 video frames and 3, 134 images. As 57% of the face images in IJB-C are natural occluded, this dataset is a commonly used benchmark for occlusion-robust face recognition. In this work, we employ two evaluation settings. First, we conduct comparisons on the holistic IJB-C dataset (including both occluded faces and non-occluded faces) for general evaluations. Second, to further verify the effectiveness of tackling occlusions, the occlusion subset of IJB-C (the occluded face images only) is employed for evaluations. All the two settings follow the standard IJB-C testing protocol. The true accept rate (TAR) and the false accept rate (FAR) are used as the evaluation metrics.
MegaFace. The MegaFace challenge 1 (MF1) benchmark [41] evaluates how the face recognition method performs with a huge scale of distracters. Specifically, the gallery set in MF1 contains one million face distractors, and the probe set Facescrub [45] contains 106,863 face images of 530 identities. In the testing pipeline, each Facescrub image will be added into the galley set and the remaining images of the same identity are exploited as probes. The rank-1 identification accuracy is used as the measurement of the face recognition performance. It should be noted that the un-cleaned version of MegaFace datasets are employed in the evaluation for the fair comparison with the state-of-the-art methods.
LFW. The LFW [42] is a well-known unconstrained face verification benchmark. It contains 13,233 images form 5,749 identities. In our work, we follow the standard 10-fold cross validation protocol to report the mean accuracy on the 6,000 testing image pairs.
V-B Implementation Details
In our experiments, the face images are aligned based on five facial landmarks detected by [46] and normalized to a size of 112112. We employ the ResNet-50 [47] as the baseline network and implement our method on the Tensorflow [48] platform. The Arcface loss [49] with the margin of 0.5 and the scale of 64 is utilized as the identification loss in our experiments. All the models in our experiments are trained on four NVIDIA TITAN XP GPUs by using SGD. The loss weight of the ArcFace loss is set to 1 and the loss weights of the filter orthogonal regularization and the response orthogonal regularization are 100 and 1, respectively.
V-C Where to deploy the Channel-wise Dropout
We firstly investigate the best stage to deploy the channel-wise dropout. By integrating the channel-wise dropout into three different stages of the plain ResNet-50, respectively, we construct experiments under three settings. Specifically, the channel-wise dropout is conducted on the outputs of the last convolution layer in the stage specific to each setting.
Methods | @FAR=.0001 | @FAR=.001 | @FAR=.01 |
---|---|---|---|
Baseline | 39.69 | 88.52 | 95.14 |
channel-wise dropout (stage 2) | 53.57 | 90.55 | 95.73 |
channel-wise dropout (stage 3) | 57.88 | 90.80 | 95.48 |
channel-wise dropout (stage 4) | 37.59 | 88.96 | 95.48 |
Table I summarizes the results on the IJB-C occlusion subset of the three different settings. As seen, the channel-wise dropout in the stage 3 outperforms the baseline by an improvement up to 18.19% in terms of TAR when FAR = 0.0001. This setting works better than the channel-wise drop conducted in the shallow stage 2. The reason behind it is that features in stage 3 have larger receptive field than those in stage 2, which can well characterize facial components, leading to better simulation of partial occlusions on faces. Besides, conducting channel-wise dropout in the stage 4 performs severely worse than that in the stage 3, or even worse than the baseline model. We argue that employing several succeeding neural layers to compensate the damaged activations plays an important role for face recognition under occlusions. We conduct a further experiment to verify this analysis as bellow.


For a face recognition model, we compare the similarity of output features from stage 4 with and without conducting channel-wise dropout in stage 3, as shown in Fig. 3. The features similarity is evaluated by mean square error (MSE) as shown in Eq. 8.
(8) |
where the denotes the length of the final feature vector and denotes the -th elements within the feature vector.
We compare the baseline mode and our model trained with the channel-wise dropout integrated in the stage 3. The results are shown in Fig. 4. As seen, our model trained with channel-wise dropout integrated in the stage 3 achieves notably lower MSE than baseline, which means the final features of our method on “occluded face” are much closer to those on clean face, leading to occlusion-robust face recognition model. The result proves that after integrating the channel-wise dropout in the stage 3, the succeeding layers in the following stage 4 play a crucial role in compensating the absent of facial information caused by occlusions. This experimental results also explain why conducting the channel-wise dropout in the stage 4 leads to poor results in Table I. If the occlusion is simulated in the last stage, no succeeding layers are enforced to learn features robust to occlusions. Based on this analysis, the channel-wise dropout integrated into the stage 3 of the ResNet-50 are treated as the best setting for all the following experiments.
V-D Ablation study
We conduct ablation studies to evaluate the effectiveness of each component in our method. The results on the IJB-C occlusion subset are shown in Table II. As seen, all the three components are beneficial to the performance improvement. Specifically, when only the channel-wise dropout is conducted, the TAR when FAR= 0.0001 is improved by 18.19%. Moreover, by jointly training with the spatial regularization, a further improvement of 6.83% is witnessed. Furthermore, by employing all the three components, our method achieves a significant improvement up to 36.58% in terms of TAR when FAR = 0.0001, which demonstrates the effectiveness of each component of our method. For simplicity, our method consisting of all the three components is abbreviated as LCD in the following experiments.
baseline | CD | SR | SAM | @FAR=.0001 | @FAR=.001 |
---|---|---|---|---|---|
✓ | 39.69 | 88.52 | |||
✓ | ✓ | 57.88 | 90.80 | ||
✓ | ✓ | ✓ | 64.71 | 91.50 | |
✓ | ✓ | ✓ | ✓ | 76.27 | 91.83 |
Methods | @FAR=.0001 | @FAR=.001 | @FAR=.01 |
---|---|---|---|
DR-GAN [50] | - | 73.6 | 88.2 |
PFE [51] | - | 89.6 | 93.3 |
DUL [52] | - | 90.2 | 94.2 |
CASIA-Net InterpretFR [35] | - | 89.2 | 75.6 |
ResNet-50 InterpretFR [35] | - | 93.2 | 95.8 |
Cutout [53] | 74.7 | 94.6 | 97.7 |
DropBlock [37] | 77.3 | 94.5 | 97.6 |
WCD [38] | 87.5 | 93.3 | 96.1 |
LCD (ours) | 89.8 | 95.4 | 97.5 |
V-E Comparisons with state-of-the-art Methods
V-E1 Evaluations on the IJB-C benchmark
Methods | @FAR=.0001 | @FAR=.001 | @FAR=.01 |
---|---|---|---|
DR-GAN [50] | - | 66.1 | 82.4 |
CASIA-Net InterpretFR [35] | - | 69.3 | 83.8 |
ResNet-50 InterpretFR [35] | - | 89.8 | 93.4 |
Cutout [53] | 47.9 | 90.5 | 96.0 |
DropBlock [37] | 51.0 | 90.4 | 95.7 |
WCD [38] | 74.6 | 88.5 | 93.1 |
LCD (ours) | 76.3 | 91.8 | 95.4 |
We firstly evaluate our LCD on the IJB-C dataset and compare it with the state-of-the-art occlusion robust method named InterpretFR [35]. The results are shown in Table III, where the result of InterpretFR is directly quoted from [35]. As seen, our LCD outperforms InterpretFR with an improvement up to 2.2% of TAR when FAR = 0.001, which demonstrates the effectiveness of our locality-aware channel-wise dropout and spatial attention module. Besides, we compare with the image augmentation method Cutout [53], which randomly puts black box on faces to simulate occlusions. Attributed to simulating more realistic occlusion by LCD, our method significantly surpasses Cutout with 15.1% improvement in terms of TAR when FAR = 0.0001. Since DropBlock [37] follows the similar spirit by randomly zeroing out continuous activations of all channels to enhance the feature representations, we conduct further comparison with DropBlock and our method also significantly outperforms it, demonstrating the superiority of proposing LCD for occlusion synthesis. Detailed analysis between our method and DropBlock are illustrated in Section IV.
To further verify the effectiveness of our LCD for tackling face recognition under occlusions, we make more challenging experiments on the IJB-C occlusion subset, which only consists of occluded faces. Similar conclusion can be achieved that our LCD outperforms InterpretFR with an improvement up to 2.0% in terms of TAR when FAR = 0.001, demonstrating the superiority of our method again. It is worth mentioning that our LCD significantly surpasses Cutout and DropBlock with improvements up to 28.4% and 25.3% in terms of TAR when FAR = 0.0001, respectively. Our LCD can simulate more realistic occlusions than both Cutout and DropBlock, which markedly improves the robustness to occlusions for face recognition. Furthermore, comparing to the weighted channel-wise dropout (WCD), our method achieves an improvement up to 3.3% of TAR when FAR = 0.001, demonstrating that our method is more superior in dealing with occluded face recognition. Detailed analysis between our method and WCD are illustrated in Section IV.
V-E2 Evaluations on the LFW benchmark
Table V summarizes the accuracy results on the LFW dataset. As seen, our LCD performs better than the occlusion discarding method PDSN [12]. Furthermore, when comparing with stronger competitor CurricularFace [54] which utilizes a larger backbone of ResNet100 and much more training images, our LCD still achieves a comparable result of 99.78%. It is worth noting that the LFW is a general face recognition benchmark which mainly consists of non-occluded faces. Therefore, these results indicates that our method also generalizes well under non-occluded scenarios.
Methods | #Image | Accuracy |
---|---|---|
DeepID [55] | 0.1M | 99.47 |
Deep Face [56] | 4.4M | 97.35 |
VGG Face [57] | 2.6M | 98.95 |
FaceNet [58] | 200M | 99.63 |
Baidu [59] | 1.3M | 99.13 |
Center Loss [60] | 0.7M | 99.28 |
Range Loss [61] | 5M | 99.52 |
Marginal Loss [62] | 3.8M | 99.48 |
SphereFace [63] | 0.5M | 99.42 |
SpherFace+ [64] | 0.5M | 99.47 |
PDSN [12] | 0.5M | 99.20 |
CosFace [65] | 5M | 99.73 |
ArcFace, R100 [49] | 5.8M | 99.83 |
MV-Arc-Softmax [66] | 3.28M | 99.78 |
CurricularFace, R100 [54] | 5.8M | 99.80 |
LCD (ours) | 3.8M | 99.78 |
V-E3 Evaluations on the MegaFace benchmark
Finally, we evaluate our method on the MegaFace which is a more challenging benchmark for general scenarios. Table VI shows the rank-1 accuracies of recent methods on this benchmark. By leveraging only 3.8 million face images for training, our method with a light backbone of ResNet-50 surpasses both ArcFace [49] and CurricularFace [54] which use ResNet-100 as a backbone and are trained with 5.8 million face images. Attributed to the LCD which encourages the network to learn more comprehensive features from faces, our method not only improves the robustness to occlusions but also enhances the feature representation learning for general face recognition.
Methods | #Image | MF1 |
---|---|---|
RegularFace [67] | 3.1M | 75.61 |
UniformFace [68] | 3.8M | 79.98 |
CosFace [65] | 5.0M | 82.72 |
ArcFace, R100 [49] | 5.8M | 81.03 |
PFE [51] | 4.4M | 78.95 |
CurricularFace, R100 [54] | 5.8M | 81.26 |
LCD (ours) | 3.8M | 83.57 |
VI Conclusion
Different from previous methods which augment face images with synthesized occlusions, we propose a novel method to better simulate realistic occlusions by dropping a group of activations in intermediate features. We first employ a spatial regularization to encourage each feature channel to respond to different face regions. In this way, the activations affected by the partial occlusion are more likely to be located in a single feature channel. Then, the locality-aware channel-wise dropout is proposed to simulate the occlusion by dropping out the entire feature channel. In addition, we design an auxiliary spatial attention module to reweight the feature channels, which can further emphasize the contributions of non-occluded regions. By directly simulating the influence of arbitrary occlusion on intermediate features, the proposed method improves the robustness against occlusion by encouraging the neural network to capture more discriminative information from the non-occluded face regions. Extensive experiments on various benchmarks have shown that the proposed method is a practical and effective approach which outperforms state-of-the-art methods with a remarkable improvement. From the practice in this work, we can conclude that the well-known dropout strategy is not only effective for improving the generalizability but also good for achieving occlusion robustness after simple modification. Our work also shows that the simulation of occlusion in feature-level rather than image-level can be a good direction to further study.
References
- [1] L. Cheng, J. Wang, Y. Gong, and Q. Hou, “Robust deep auto-encoder for occluded face recognition,” in ACM international conference on Multimedia, 2015, pp. 1099–1102.
- [2] F. Zhao, J. Feng, J. Zhao, W. Yang, and S. Yan, “Robust lstm-autoencoders for face de-occlusion in the wild,” IEEE Transactions on Image Processing, vol. 27, no. 2, pp. 778–790, 2017.
- [3] D. Pathak, P. Krahenbuhl, J. Donahue, T. Darrell, and A. A. Efros, “Context encoders: Feature learning by inpainting,” in IEEE conference on computer vision and pattern recognition (CVPR), 2016, pp. 2536–2544.
- [4] S. Iizuka, E. Simo-Serra, and H. Ishikawa, “Globally and locally consistent image completion,” ACM Transactions on Graphics (ToG), vol. 36, no. 4, pp. 1–14, 2017.
- [5] Y. Li, S. Liu, J. Yang, and M.-H. Yang, “Generative face completion,” in IEEE conference on computer vision and pattern recognition (CVPR), 2017, pp. 3911–3919.
- [6] X. Yuan and I. K. Park, “Face de-occlusion using 3d morphable model and generative adversarial network,” in IEEE International Conference on Computer Vision (ICCV), 2019, pp. 10 062–10 071.
- [7] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in Advances in neural information processing systems, 2014, pp. 2672–2680.
- [8] R. Min, A. Hadid, and J.-L. Dugelay, “Improving the recognition of faces occluded by facial accessories,” in IEEE International Conference on Automatic Face & Gesture Recognition (FG), 2011, pp. 442–447.
- [9] Z. Chen, T. Xu, and Z. Han, “Occluded face recognition based on the improved svm and block weighted lbp,” in International Conference on Image Analysis and Signal Processing. IEEE, 2011, pp. 118–122.
- [10] S. Park, H. Lee, J.-H. Yoo, G. Kim, and S. Kim, “Partially occluded facial image retrieval based on a similarity measurement,” Mathematical Problems in Engineering, vol. 2015, 2015.
- [11] W. Wan and J. Chen, “Occlusion robust face recognition based on mask learning,” in 2017 IEEE International Conference on Image Processing (ICIP), 2017, pp. 3795–3799.
- [12] L. Song, D. Gong, Z. Li, C. Liu, and W. Liu, “Occlusion robust face recognition based on mask learning with pairwise differential siamese network,” in IEEE International Conference on Computer Vision (ICCV), 2019, pp. 773–782.
- [13] F. Ding, P. Peng, Y. Huang, M. Geng, and Y. Tian, “Masked face recognition with latent part detection,” in ACM International Conference on Multimedia, 2020, pp. 2281–2289.
- [14] J.-J. Lv, X.-H. Shao, J.-S. Huang, X.-D. Zhou, and X. Zhou, “Data augmentation for face recognition,” Neurocomputing, vol. 230, pp. 184–196, 2017.
- [15] E. Osherov and M. Lindenbaum, “Increasing cnn robustness to occlusions by reducing filter support,” in IEEE International Conference on Computer Vision (ICCV), 2017, pp. 550–561.
- [16] D. S. Trigueros, L. Meng, and M. Hartnett, “Enhancing convolutional neural networks for face recognition with occlusion maps and batch triplet loss,” Image and Vision Computing, vol. 79, pp. 99–108, 2018.
- [17] C. Shao, J. Huo, L. Qi, Z.-H. Feng, W. Li, C. Dong, and Y. Gao, “Biased feature learning for occlusion invariant face recognition,” in International Joint Conference on Artificial Intelligence (IJCAI), 2020, pp. 666–672.
- [18] D. Novotny, D. Larlus, and A. Vedaldi, “Anchornet: A weakly supervised network to learn geometry-sensitive features for semantic matching,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
- [19] J.-S. Park, Y. H. Oh, S. C. Ahn, and S.-W. Lee, “Glasses removal from facial image using recursive error compensation,” IEEE transactions on pattern analysis and machine intelligence, vol. 27, no. 5, pp. 805–811, 2005.
- [20] S. Fidler, D. Skocaj, and A. Leonardis, “Combining reconstructive and discriminative subspace methods for robust classification and regression by subsampling,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, no. 3, pp. 337–350, 2006.
- [21] E. J. Candès, X. Li, Y. Ma, and J. Wright, “Robust principal component analysis,” Journal of the ACM (JACM), vol. 58, no. 3, pp. 1–37, 2011.
- [22] G. Liu, Z. Lin, S. Yan, J. Sun, Y. Yu, and Y. Ma, “Robust recovery of subspace structures by low-rank representation,” IEEE transactions on pattern analysis and machine intelligence, vol. 35, no. 1, pp. 171–184, 2012.
- [23] F. Zhang, J. Yang, Y. Tai, and J. Tang, “Double nuclear norm-based matrix decomposition for occluded image recovery and background modeling,” IEEE Transactions on Image Processing, vol. 24, no. 6, pp. 1956–1966, 2015.
- [24] Y. Qian, W. Deng, and J. Hu, “Unsupervised face normalization with extreme pose and expression in the wild,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 9851–9858.
- [25] H. J. Oh, K. M. Lee, and S. U. Lee, “Occlusion invariant face recognition using selective local non-negative matrix factorization basis images,” Image and Vision computing, vol. 26, no. 11, pp. 1515–1523, 2008.
- [26] J. Hu, J. Lu, and Y.-P. Tan, “Robust partial face recognition using instance-to-class distance,” in 2013 Visual Communications and Image Processing (VCIP), 2013, pp. 1–6.
- [27] L. He, H. Li, Q. Zhang, and Z. Sun, “Dynamic feature matching for partial face recognition,” IEEE Transactions on Image Processing, vol. 28, no. 2, pp. 791–802, 2018.
- [28] J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, “Robust face recognition via sparse representation,” IEEE transactions on pattern analysis and machine intelligence, vol. 31, no. 2, pp. 210–227, 2008.
- [29] J. Dong, H. Zheng, and L. Lian, “Low-rank laplacian-uniform mixed model for robust face recognition,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 11 897–11 906.
- [30] M. Yang, L. Zhang, S. C. Shiu, and D. Zhang, “Gabor feature based robust representation and classification for face recognition with gabor occlusion dictionary,” Pattern Recognition, vol. 46, no. 7, pp. 1865–1878, 2013.
- [31] K. Jia, T.-H. Chan, and Y. Ma, “Robust and practical face recognition via structured sparsity,” in European conference on computer vision (ECCV), 2012, pp. 331–344.
- [32] M. Yang, X. Wang, G. Zeng, and L. Shen, “Joint and collaborative representation with local adaptive convolution feature for face recognition with single sample per person,” Pattern recognition, vol. 66, pp. 117–128, 2017.
- [33] W. Li, X. Zhu, and S. Gong, “Harmonious attention network for person re-identification,” in IEEE conference on computer vision and pattern recognition (CVPR), 2018, pp. 2285–2294.
- [34] K. Wang, X. Peng, J. Yang, D. Meng, and Y. Qiao, “Region attention networks for pose and occlusion robust facial expression recognition,” IEEE Transactions on Image Processing, vol. 29, pp. 4057–4069, 2020.
- [35] B. Yin, L. Tran, H. Li, X. Shen, and X. Liu, “Towards interpretable face recognition,” in IEEE International Conference on Computer Vision (ICCV), 2019, pp. 9348–9357.
- [36] J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” in IEEE conference on computer vision and pattern recognition (CVPR), 2018, pp. 7132–7141.
- [37] G. Ghiasi, T.-Y. Lin, and Q. V. Le, “Dropblock: A regularization method for convolutional networks,” in Advances in Neural Information Processing Systems, 2018, pp. 10 727–10 737.
- [38] S. Hou and Z. Wang, “Weighted channel dropout for regularization of deep convolutional neural network,” in AAAI Conference on Artificial Intelligence, 2019, pp. 8425–8432.
- [39] Y. Guo, L. Zhang, Y. Hu, X. He, and J. Gao, “MS-Celeb-1M: A dataset and benchmark for large scale face recognition,” in European Conference on Computer Vision (ECCV), 2016, pp. 87–102.
- [40] B. Maze, J. Adams, J. A. Duncan, N. Kalka, T. Miller, C. Otto, A. K. Jain, W. T. Niggel, J. Anderson, J. Cheney, and P. Grother, “Iarpa janus benchmark-c: Face dataset and protocol,” in International Conference on Biometrics (ICB), 2018, pp. 158–165.
- [41] I. Kemelmacher-Shlizerman, S. M. Seitz, D. Miller, and E. Brossard, “The megaface benchmark: 1 million faces for recognition at scale,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 4873–4882.
- [42] G. B. Huang, M. Ramesh, T. Berg, and E. Learned-Miller, “Labeled faces in the wild: A database for studying face recognition in unconstrained environments,” University of Massachusetts, Amherst, Tech. Rep. 07-49, October 2007.
- [43] B. F. Klare, B. Klein, E. Taborsky, A. Blanton, J. Cheney, K. Allen, P. Grother, A. Mah, M. Burge, and A. K. Jain, “Pushing the frontiers of unconstrained face detection and recognition: Iarpa janus benchmark a,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015, pp. 1931–1939.
- [44] C. Whitelam, E. Taborsky, A. Blanton, B. Maze, J. Adams, T. Miller, N. Kalka, A. K. Jain, J. A. Duncan, K. Allen, J. Cheney, and P. Grother, “Iarpa janus benchmark-b face dataset,” in IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2017, pp. 592–600.
- [45] H.-W. Ng and S. Winkler, “A data-driven approach to cleaning large face datasets,” in IEEE International Conference on Image Processing (ICIP), 2014, pp. 343–347.
- [46] Z. He, M. Kan, J. Zhang, X. Chen, and S. Shan, “A fully end-to-end cascaded cnn for facial landmark detection,” in IEEE International Conference on Automatic Face & Gesture Recognition (FG), 2017, pp. 200–207.
- [47] K. He, X. Zhang, S. Ren, and J. Sun, “Identity mappings in deep residual networks,” in European Conference on Computer Vision (ECCV), 2016, pp. 630–645.
- [48] M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis, J. Dean, M. Devin, S. Ghemawat, G. Irving, M. Isard, M. Kudlur, J. Levenberg, R. Monga, S. Moore, D. G. Murray, B. Steiner, P. Tucker, V. Vasudevan, P. Warden, M. Wicke, Y. Yu, and X. Zheng, “Tensorflow: A system for large-scale machine learning,” in USENIX Symposium on Operating Systems Design and Implementation (OSDI), 2016, pp. 265–283.
- [49] J. Deng, J. Guo, N. Xue, and S. Zafeiriou, “Arcface: Additive angular margin loss for deep face recognition,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 4690–4699.
- [50] L. Tran, X. Yin, and X. Liu, “Representation learning by rotating your faces,” IEEE transactions on pattern analysis and machine intelligence, vol. 41, no. 12, pp. 3007–3021, 2018.
- [51] Y. Shi and A. K. Jain, “Probabilistic face embeddings,” in IEEE International Conference on Computer Vision (ICCV), 2019, pp. 6902–6911.
- [52] J. Chang, Z. Lan, C. Cheng, and Y. Wei, “Data uncertainty learning in face recognition,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 5710–5719.
- [53] T. DeVries and G. W. Taylor, “Improved regularization of convolutional neural networks with cutout,” arXiv preprint arXiv:1708.04552, 2017.
- [54] Y. Huang, Y. Wang, Y. Tai, X. Liu, P. Shen, S. Li, J. Li, and F. Huang, “Curricularface: adaptive curriculum learning loss for deep face recognition,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 5901–5910.
- [55] Y. Sun, Y. Chen, X. Wang, and X. Tang, “Deep learning face representation by joint identification-verification,” in Advances in neural information processing systems, 2014, pp. 1988–1996.
- [56] Y. Taigman, M. Yang, M. Ranzato, and L. Wolf, “Deepface: Closing the gap to human-level performance in face verification,” in IEEE conference on computer vision and pattern recognition (CVPR), 2014, pp. 1701–1708.
- [57] O. M. Parkhi, A. Vedaldi, and A. Zisserman, “Deep face recognition,” in British Machine Vision Association (BMVC), 2015.
- [58] F. Schroff, D. Kalenichenko, and J. Philbin, “Facenet: A unified embedding for face recognition and clustering,” in IEEE conference on computer vision and pattern recognition (CVPR), 2015, pp. 815–823.
- [59] J. Liu, Y. Deng, T. Bai, Z. Wei, and C. Huang, “Targeting ultimate accuracy: Face recognition via deep embedding,” arXiv preprint arXiv:1506.07310, 2015.
- [60] Y. Wen, K. Zhang, Z. Li, and Y. Qiao, “A discriminative feature learning approach for deep face recognition,” in European conference on computer vision (ECCV), 2016, pp. 499–515.
- [61] X. Zhang, Z. Fang, Y. Wen, Z. Li, and Y. Qiao, “Range loss for deep face recognition with long-tailed training data,” in IEEE International Conference on Computer Vision (ICCV), 2017, pp. 5409–5418.
- [62] J. Deng, Y. Zhou, and S. Zafeiriou, “Marginal loss for deep face recognition,” in IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2017, pp. 60–68.
- [63] W. Liu, Y. Wen, Z. Yu, M. Li, B. Raj, and L. Song, “Sphereface: Deep hypersphere embedding for face recognition,” in IEEE conference on computer vision and pattern recognition (CVPR), 2017, pp. 212–220.
- [64] W. Liu, R. Lin, Z. Liu, L. Liu, Z. Yu, B. Dai, and L. Song, “Learning towards minimum hyperspherical energy,” Advances in neural information processing systems, vol. 31, pp. 6222–6233, 2018.
- [65] H. Wang, Y. Wang, Z. Zhou, X. Ji, D. Gong, J. Zhou, Z. Li, and W. Liu, “Cosface: Large margin cosine loss for deep face recognition,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018, pp. 5265–5274.
- [66] X. Wang, S. Zhang, S. Wang, T. Fu, H. Shi, and T. Mei, “Mis-classified vector guided softmax loss for face recognition,” in AAAI Conference on Artificial Intelligence, vol. 34, no. 07, 2020, pp. 12 241–12 248.
- [67] K. Zhao, J. Xu, and M.-M. Cheng, “Regularface: Deep face recognition via exclusive regularization,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 1136–1144.
- [68] Y. Duan, J. Lu, and J. Zhou, “Uniformface: Learning deep equidistributed representation for face recognition,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 3415–3424.
![]() |
Mingjie He received the M.S. degree from the University of Science and Technology of China, Hefei, China, in 2014. Currently, he is a Ph.D. candidate at the University of Chinese Academy of Sciences and an engineer with the Institute of Computing Technology, Chinese Academy of Sciences (CAS). His research interests cover computer vision and machine learning. |
![]() |
Jie Zhang is an associate professor with the Institute of Computing Technology, Chinese Academy of Sciences (CAS). He received the Ph.D. degree from the University of Chinese Academy of Sciences, Beijing, China. His research interests cover computer vision, pattern recognition, machine learning, particularly include face recognition, image segmentation, weakly/semi-supervised learning, domain generalization. |
![]() |
Shiguang Shan received Ph.D. degree in computer science from the Institute of Computing Technology (ICT), Chinese Academy of Sciences (CAS), Beijing, China, in 2004. He has been a full Professor of this institute since 2010 and now the deputy director of CAS Key Lab of Intelligent Information Processing. His research interests cover computer vision, pattern recognition, and machine learning. He has published more than 300 papers, with totally more than 15,000 Google scholar citations. He has served as Area Chair (or Senior PC) for many international conferences including ICCV11, ICPR12/14/20, ACCV12/16/18, FG13/18, ICASSP14, BTAS18, AAAI20/21, IJCAI21, and CVPR19/20/21. And he was/is Associate Editors of several journals including IEEE T-IP, Neurocomputing, CVIU, and PRL. He was a recipient of the China’s State Natural Science Award in 2015, and the China’s State S&T Progress Award in 2005 for his research work. |
![]() |
Xiao Liu is a researcher with Tomorrow Advancing Life Education Group (TAL). He received the PhD degree in computer science from Zhejiang University in 2015 and then worked for Baidu from 2015 to 2019. His research interests is the applied AI such as intelligent multimedia processing, computer vision and learning system. His research results have expounded in 40+ publications at journals and conferences such as IEEE T-IP, T-NNLS, CVPR, ICCV, ECCV, AAAI and MM. As a key team member, he achieved the best performance in various competitions, such as the ActivityNet challenges, NTIRE super resolution challenge, EmotioNet facial expression recognition challenge, etc. |
![]() |
Zhongqin Wu is a scientist in the Tomorrow Advancing Life Education Group (TAL) and in charge of the AI laboratory. He received the master degree in computer science from Fudan University, China. He previously worked for Baidu and leaded the Department of Computer Vision and Augmented Reality Laboratory. |
![]() |
Xilin Chen is a professor with the Institute of Computing Technology, Chinese Academy of Sciences (CAS). He has authored one book and more than 200 papers in refereed journals and proceedings in the areas of computer vision, pattern recognition, image processing, and multimodal interfaces. He was a recipient of several awards, including the China’s State Natural Science Award in 2015, the China’s State S&T Progress Award in 2000, 2003, 2005, and 2012 for his research work. He is currently an associate editor of the IEEE Transactions on Multimedia, and Journal of Visual Communication and Image Representation, a leading editor of the Journal of Computer Science and Technology, and an associate editor-in-chief of the Chinese Journal of Computers, and Chinese Journal of Pattern Recognition and Artificial Intelligence. He served as an Organizing Committee member for many conferences, including general co-chair of FG13/FG18, Local chair of ICME07, ACM MM09, and ICIP17, and Finance chair of ISCAS13. He is/was an area chair of CVPR 2017/2019/2020, and ICCV 2019. He is a fellow of the ACM, IEEE, IAPR, and CCF. |