Learning Facial Representations from the Cycle-consistency of Face
Abstract
Faces manifest large variations in many aspects, such as identity, expression, pose, and face styling. Therefore, it is a great challenge to disentangle and extract these characteristics from facial images, especially in an unsupervised manner. In this work, we introduce cycle-consistency in facial characteristics as free supervisory signal to learn facial representations from unlabeled facial images. The learning is realized by superimposing the facial motion cycle-consistency and identity cycle-consistency constraints. The main idea of the facial motion cycle-consistency is that, given a face with expression, we can perform de-expression to a neutral face via the removal of facial motion and further perform re-expression to reconstruct back to the original face. The main idea of the identity cycle-consistency is to exploit both de-identity into mean face by depriving the given neutral face of its identity via feature re-normalization and re-identity into neutral face by adding the personal attributes to the mean face. At training time, our model learns to disentangle two distinct facial representations to be useful for performing cycle-consistent face reconstruction. At test time, we use the linear protocol scheme for evaluating facial representations on various tasks, including facial expression recognition and head pose regression. We also can directly apply the learnt facial representations to person recognition, frontalization and image-to-image translation. Our experiments show that the results of our approach is competitive with those of existing methods, demonstrating the rich and unique information embedded in the disentangled representations. Code is available at https://github.com/JiaRenChang/FaceCycle.
1 Introduction

Face perception is vital for human beings and is also essential in the field of computer vision. Neuroimaging studies of both human and monkey [13, 15, 43] reveal the neuroanatomical dissociation between expression and identity representations in face perception. Their findings suggest that these facial characteristics are processed in different brain areas. With the renaissance of deep learning in recent years, computer vision research field follows this thread of thinking and progresses in the direction of disentangling the face characteristics into separated low-dimensional latent representations, such as identity [40], expression [45, 48], shape/appearance [35, 44], intrinsic images [36], and fine-grained attributes (age, gender, wearing glasses, etc.) [34].
Several supervised methods have been proposed to disentangle face characteristics for image manipulation by conditioning generative models on a pre-specified face representations, including landmarks [47], action units [32], or facial attributes [27]. Particularly, these methods are able to manipulate faces while preserving the identity. Other studies incorporate head pose information to disentangle pose-invariant representations for robust identity [40]/expression [48] recognition. Provided with neutral face, moreover, de-expression residue learning [45] can facilitate the model to learn identity-invariant expression representations for performing facial expression recognition.
The 3D Morphable Model (3DMM) [2, 4] for face shape modeling incorporates a similar thinking of dissociation for expression and identity. The most widely-used form of 3DMM is that a face shape is a linear combination of mean shape and identity and expression vectors (): , where and are the identity and expression PCA bases, respectively. Jiang et al. [20] introduce a variational autoencoder approach for learning latent representations of expression mesh and identity mesh in the framework of 3DMM. However, they provided strong supervision for the disentanglement of identity and expression representations, including ground truths of shape meshes for expression, identity, and the mean face [20]. It is difficult to generalize such methods to 2D facial images without being given any ground truth.
In addition to the aforementioned works which are mostly based on supervised learning, recently a few studies begin to exploit the unsupervised learning framework to disentangle facial characteristics [26, 41, 42, 44]. These methods focus on extracting a part of facial characteristics. For example, FAb-Net [41] learns representations that encode information about pose and expression, [26, 41] introduce frameworks to learn representations for action unit detection, and Zhang et al. [49] propose an autoencoder to locate facial landmarks. Some unsupervised methods [35, 44] attempt to separate two independent representations of face images, including shape and appearance. However, these unsupervised methods can only disentangle a part of information of facial images, but not yet investigate a more general generative procedure of a human face, that is, simultaneous disentanglement of expression and identity representations for a wider usage.
In this paper, we propose a novel framework that is able to simultaneously disentangle expression and identity representations from 2D facial images in an unsupervised manner. Particularly, the definition of the expression factor in our proposed method contains all the variations between an arbitrary face image and its corresponding neutral face of the same identity, including the facial expression and head pose. While for the identity factor, we define it to contain all the variations between a neutral face and the global mean face, including the facial identity and other subject-specific features such as hair style, age, gender, beard, glasses, etc. Based on these definitions, we propose two novel cycle-consistency constraints to drive our model learning, as illustrated in Figure 1.
The first cycle-consistency constraint stems from the idea of action unit [9] in which the head poses and facial expressions are the results of the combined and coordinated action of facial muscles. Therefore, the head poses and expression can be treated as the optical flow [28] between a neutral face and any face of the same identity. To this end, a decoder is trained to learn the optical flow field of the input face without the ground truth neutral face. This is achieved by applying the proposed idea called facial motion cycle-consistency, which is able to perform both the de-expression and re-expression operations.
The second cycle-consistency constraint originates from Eigenfaces [38], in which a facial image is represented by adding a linear combination of eigenfaces to the mean face, suggesting that the face identity is embedded in the linear combination of eigenfaces. Instead of representing the identity as the residue of neutral facial image relative to the mean face [38], we model the adding and depriving of identity as a renormalization procedure, analogues to the feed-forward style transfer tasks [18]. To this end, decoders are trained to learn the renormalized features without the ground truth mean face. This is achieved by applying the proposed idea called identity cycle-consistency, which is able to perform identity deprivation as de-identity and the identity styling as re-identity.
The main contributions of our work are summarized as follows:
-
•
We propose a novel framework for unsupervised learning of facial representations from a single facial image, based on the novel ideas of facial motion cycle-consistency and identity cycle-consistency.
-
•
The disentangled expression and identity features obtained by our proposed method can be easily utilized for various downstream tasks, such as facial expression recognition, head pose regression, person recognition, frontalization, and image-to-image translation.
-
•
We demonstrate that the performance of the learned representations in different downstream tasks is competitive with the state-of-the-art methods.
2 Unsupervised Learning of Facial Representations

As motivated previously, in this paper we aim at disentangling the identity and expression representations from a single facial image. Our proposed method is mainly based on an important assumption that: a facial image , from high-level perspective, can be decomposed as follows:
(1) |
where is the global mean face shared among all the faces, id and exp are the identity and expression factors respectively, and is the neutral face of a particular identity specified by id. Therefore, our proposed model is trained to learn the expression and identity representations, denoted as and respectively, for indicating the facial characteristics of facial images. We introduce four processes based on cycle-consistency for learning these representations, as shown in Figure 1:
-
•
de-expression. We define the de-expression as removing from the input facial image , in which we can obtain the neutral face accordingly.
-
•
re-expression. The re-expression is defined as assigning to the neutral face for reconstructing face with expressions .
-
•
de-identity. We define de-identity as an operation for removing from the input neutral face in order to obtain the mean face .
-
•
re-identity. The re-identity is defined as the process of recovering the neutral face back from the mean face according to .
As illustrated in Figure 2, the overall architecture of our proposed model consists of two encoders ( and ) for extracting expression and identity representations respectively, and two decoders (, ) for learning nonlinear mapping functions of the aforementioned four processes. In the following, we detail the proposed unsupervised learning method to disentangle expression and identity representations.

2.1 Expression Representation
We start with the introduction of facial motion cycle-consistency in the following. We denote that the expression representations are learned by an encoder from the input face image :
(2) |
As the idea described in the previous section, we model a facial expression as the optical flow field between the neutral face and the face with expression. Therefore, the forward () optical flow field , where and is the height and width of the input image, is learned by the decoder from expression representations. Moreover, according to the well-known forward-backward flow consistency [1, 19], we can compute the backward () optical flow field according to , which basically is inverse by a warp function :
(3) |
We use bilinear interpolation to implement the warping operation as in [39]. By using the forward optical flow field we can warp pixel-wisely to obtain an intermediate facial image, denoted as . Followed by using the corresponding backward optical flow field , we are able to warp back from to reconstruct . This procedure straightforwardly lead to a reconstruction loss which is defined as:
(4) |
Furthermore, we exploit a general image feature extraction to represent a face image, that is, the coarse-to-fine feature maps obtained from layers conv21, and conv31 of VGG19 network pre-trained on ImageNet [37]. Given a forward flow field , we simply use the bilinear interpolation function to obtain of the size equal to . The de-expression is then achieved by first warping with and then adopting a decoder to generate neutral face image :
(5) |
Moreover, we argue that the image features of a neutral face obtained by VGG19 could be warped back via the downsampled backward flow and then be fed into the decoder for reconstructing a face with expression, denoted as , which ideally should be identical to the original face . This process is exactly the re-expression:
(6) |
2.2 Facial Motion Cycle-Consistency: Invariance for Learning Expression Representation
The change on a face image caused by a facial motion can be expressed in terms of a spatial image transformation , where we denote the corresponding face image with different motion but the same identity as . As both and are with the same identity, their corresponding neutral faces should be also identical. That is, their decoded neutral faces after performing de-expression are invariant to each other, which leads to the constraint:
(7) |
Following the concept of this invariance, we should be able to apply the re-expression operation on (the features for the decoded neutral face of ) via the downsampled backward flow of (related to the expression of ) to reconstruct a face image denoted as , which ideally is quite similar to the original due to the hypothesis that as . The similar story holds for performing re-expression on by to reconstruct , which is almost identical to . The illustration of this invariance, also named as facial motion cycle-consistency, is shown in Figure 3(a).
The reconstruction derived from the invariance (that is versus and versus ) builds up the objectives for learning the expression representations , where we utilize both the L1 loss and the perceptual loss [11, 21] to evaluate the error of reconstruction:
(8) | ||||
where is set to to balance the L1 and perceptual losses. The perceptual loss is defined as . The function extracts VGG19 features from layer , in which the conv2_1, conv3_1, and conv4_1 layers are used here. The function calculates the Gram matrix of the feature map.
2.3 Identity Representation
In terms of identity representation , we utilize the encoder to extract from the input face image :
(9) |
Based on the idea described previously, we argue that the identity representation could be deprived from the neutral face to obtain the mean face. To implement the de-identity operation, we design a decoder to generate the mean face from the modulated VGG features of a neutral face , which is similar to the idea of feature modulation idea proposed in the AdaIN paper [18]:
(10) |
where and are used to compute the mean and standard deviation respectively, and and are learned from by the multi-layer perceptron :
(11) |
Furthermore, the re-identity can be achieved in a similar manner but reversely with the decoder :
(12) |
where the and are also learned from but by another multi-layer perceptron .
2.4 Identity Cycle-Consistency: Invariance for Learning Identity Representation
We hypothesize that the mean face is global for all the faces. In other words, no matter starting from which neural face of any identity, we should always obtain the same mean face after performing de-identity operation. Given the neutral faces and of different identities, we can derive the invariance related to identity as:
(13) |
Therefore, we should be able to reconstruct by using its corresponding to apply the re-identity operation on the mean face obtained from . The result of this reconstruction is denoted as :
(14) |
Again, similar story holds to perform re-identity operation (with ) on the mean face obtained from to reconstruct . We denote the reconstruction result as :
(15) |
The illustration of this invariance related to identity representations, also named as identity cycle-consistency, is shown in Figure 3(b).
As the way that is defined, the reconstruction derived from the invariance (that is, versus and versus ) leads to the objectives for learning the identity representations :
(16) | ||||
Moreover, we additionally introduce a margin loss to constrain the mean face:
(17) |
where we set in all experiments. The main motivation behind this margin loss is that we would like to constrain the difference between the mean face and the neutral face to be within a margin. Otherwise the obtained mean face could potentially become an arbitrary image far from a face image.
3 Experiments
We report experimental results for a model trained on the combination of VoxCeleb1 [29] and VoxCeleb2 [5] from scratch. The trained representations are evaluated on several tasks, including facial expression recognition, head pose regression, person recognition, frontalization, and image-to-image translation. Through various experiments, we show that the acquired representation generalizes to a range of facial image processing tasks.
3.1 Training Procedures
The facial motion cycle-consistency described in Section 2.2 involves an image pair of faces with different expressions/poses but of the same identity. Fortunately, this type of data can be easily available from the video recording of human faces, for instance, the video of interview or talk-show, which exists widely on the Internet nowadays. Given any two frames in this type of video clip of a person, we can easily obtain a pair of facial images showing different expressions. Therefore, we can take the advantage of this type of video sequences (as the dataset described in the following) and collect training data for learning both the expression and identity representations in an unsupervised manner.
Dataset.
The proposed model is trained on the combination of VoxCeleb1 [29] and VoxCeleb2 [5] datasets, in which both datasets are built upon videos of interviews. VoxCeleb1 has in total 153,516 video clips of 1,251 speakers, while VoxCeleb2 has 145,569 video clips of 5,994 speakers. Video frames were extracted at 6 fps, cropped to have faces shown in the center of frames, and then resized to the resolution of 6464. We adopted VoxCeleb2 test dataset for visualizing the intermediate results of our disentanglement process.
Stage-wise Training Procedure.
We introduce a stage-wise training procedure for our model learning. There are two main stages for sequentially training different parts of the proposed model, in order to disentangle the expression and identity representations.
– Stage 1: training of , and
For training the subnetworks related to the de-expression and re-expression parts, as the green-shaded components shown in Figure 2, the objectives of and are utilized to update .
The transformation needed for the use of can be simply obtained by having the horizontal flipping (that is, is the horizontally flipped version of ) or taking any arbitrary pair of faces from different frames (that is, two faces of the same person shown at different times in a video).
We provide an ablation study in the supplementary materials.
– Stage 2: training of , , , and
For training the subnetworks related to the de-identity and re-identity parts, as the orange-shaded components shown in Figure 2, both the objectives of and are applied to update all of these subnetworks.
Implementation Details.
Our proposed model is implemented based on PyTorch framework and trained with the Adam optimizer (, and ). The batch size is set to 32 for all the training stages. The initial learning rate is 0.00005 in the Stage 1 and 0.0001 in the Stage 2. The Stage 1 and Stage 2 are trained for 40 and 20 epochs respectively. The learning rate is decreased by a factor of 10 at half of total epochs. Moreover, both representation encoders (i.e. and ) adopt the same network architecture which is a 16-layer CNN. We leverage a VGG-19 [37] for the general feature extraction (denoted as VGG19 component in the Figure 2), where the VGG-19 encoded facial features can be further passed through our decoders (i.e. or ) to generate new facial images. The model architectures are detailed in the supplementary materials.
Baselines.
We adopt the following baselines for making evaluations and comparisons in terms of the quality and representativeness of the extracted facial features:
– HoG descriptor [6]:
We follow the same setting as in [23], where the facial images are first rescaled to the size of 100100, then the HoG feature of 3,240 dimensions is extracted for each image.
– LBP descriptor [30]:
Similar to the HoG descriptor, we follow the same setting as in [23] to extract 1,450 dimensional LBP feature vector from each of the facial images which are resized to 100100.
– MoCo [16]:
We adopt the state-of-the-art self-supervised representation learning method, MoCo, as a strong baseline for us to compare with. We follow the MoCo algorithm to train the feature extractor (in which its network architecture is the same as our encoders) based on the same training dataset as ours (i.e. VoxCeleb1 and VoxCeleb2). The training runs for 40 epochs with SGD optimizer, batch size of 128, momentum 0.999, and 65,536 negative keys.
– Self-supervisely learnt facial representations:
Three state-of-the-art self-supervised frameworks of facial representation learning [24, 26, 41] are utilized to compare with our work.
We directly adopt the models officially released by their authors (which are all pretrained on the Voxceleb dataset) for experimenting the downstream tasks of expression classification and head pose regression.
Please note that we apply a linear protocol on their learnt features to have a fair comparison.
3.2 Intermediate Results of Our Model
Figure 4 illustrates several examples of the intermediate results obtained from our model, including the input faces, forward flow fields, neutral faces, backward flow fields, mean faces, and the faces reconstructed from their neutral ones. We demonstrate that the proposed method can handle face images with large variation in poses and can preserve facial attributes such as wearing glasses or beard.
Visualization of the facial motion flows presents both the head motion and the movement of facial muscles. The neutral faces are deprived of facial motions in comparison to their original facial images. Moreover, the mean faces obtained from different input images are almost identical to each other, which is in line with our assumption of identity invariance.

3.3 Evaluation for Expression Representation
Given the trained model, we investigate the learnt expression representation by evaluating the performance of its applications on expression recognition and head pose regression. The goal is to verify whether the expression representation successfully encodes the information related to the facial motions and poses as our definition (i.e. the expression factor contains all the variations between a face image and its corresponding neutral face of the same identity, including facial motions and head poses). We conduct linear-protocol evaluation scheme for demonstrating the effectiveness of our method.
3.3.1 Expression Recognition
Two datasets are used in the experiments of expression recognition, i.e. FER-2013 [12] and RAF-DB [23]. FER-2013 dataset [12] consists of 28,709 training and 3,589 testing images, while RAF-DB dataset consists of around 30K diverse facial images downloaded from the Internet. Please note that for the RAF-DB dataset, we follow the experimental setup as [23] to particularly use the basic emotion subset of RAF-DB, which includes 12,271 training and 3,068 testing images. For the evaluation scheme of linear-protocol, in order to directly verify the capacity of the expression features extracted by different models, we construct the linear classifier upon the frozen expression representations to perform the expression recognition, as in [16]. We follow the same procedure as [16] to train the linear layer (as the classifier) for 300 epochs, where the learning rate starts from 30 and decreases by a factor of 10 for every 80 epochs. The classifiers are trained by the SGD optimizer with cross-entropy objective and 256 batch size.
The quantitative results shown in Table 1 demonstrate that the expression representation extracted from our proposed method is able to provide superior performance with respect to all the baselines. These results suggest that our proposed method can be used as a pretext task for expression recognition, where the rich information of facial expression is well learnt in a self-supervised manner.
FER-2013 | RAF-DB | |
Method | Accuracy () | Accuracy () |
Fully supervised | ||
FSN [50] | 67.60 | 81.10 |
ALT [10] | 69.85 | 84.50 |
Linear classification protocol | ||
LBP | 37.89 | 52.17 |
HoG | 45.47 | 63.53 |
FAb-Net [41] | 46.98 | 66.72 |
TCAE [24] | 45.05 | 65.32 |
BMVC’20 [26] | 47.61 | 58.86 |
MoCo | 47.24 | 68.32 |
Ours | 48.76 | 71.01 |
3.3.2 Regression of Head Pose
Our definition indicates that the information of head poses would be also encoded into the expression representations. Obviously the calculated flow fields using the proposed method contain not only the local facial motion but also the global head motion, suggesting that our expression representation can also be used in the task of head pose regression. We adopt the 300W-LP [33] dataset and the AFLW2000 [52] dataset as the training and testing sets respectively, for experimenting the head pose regression. For the evaluation scheme of linear-protocol, we construct a linear regressor on top of the frozen expression representations . The training runs for 300 epochs for the inear-protocol with SGD optimizer and batch size set to 16.
As shown in Table 2, for the linear-protocol evaluation scheme, the regressor based on our expression representations achieves 12.47 in terms of mean absolute error (MAE), which outperforms all the baselines. These results demonstrate the effectiveness of our proposed method for well capturing the head pose information into expression representations.
Method | Yaw | Pitch | Roll | MAE |
---|---|---|---|---|
Fully supervised | ||||
FAN [3] | 6.36 | 12.3 | 8.71 | 9.12 |
FSA-Net [46] | 5.27 | 6.71 | 5.28 | 5.75 |
Linear regression protocol | ||||
Dlib (68 points) [22] | 23.10 | 13.60 | 10.50 | 15.80 |
LBP | 23.58 | 14.86 | 16.36 | 18.27 |
HoG | 13.94 | 13.17 | 14.92 | 14.00 |
FAb-Net [41] | 13.92 | 13.25 | 14.51 | 13.89 |
TCAE [24] | 21.75 | 14.57 | 14.83 | 17.39 |
BMVC’20 [26] | 22.06 | 13.50 | 15.14 | 16.90 |
MoCo | 28.49 | 16.29 | 15.55 | 20.11 |
Ours | 11.70 | 12.76 | 12.94 | 12.47 |
3.4 Evaluating Identity Representations
We also investigate the applications of identity representations learned by using the proposed method on the VoxCeleb dataset. Good performance of person recognition demonstrates that our identity representations do contain rich information related to identities.
3.4.1 Person Recognition
In this work we adopt LFW [17] and CPLFW [51] dataset for the evaluation of person recognition, particularly on person verification. The LFW dataset comprises of 13,233 face images from 5,749 identities and has 6,000 face pairs for evaluating person verification. The CPLFW dataset is similar to LFW but includes larger head pose variation. We directly extract the identity representations for all of the images in the face pairs from two datasets by using the encoder and then compute the cosine similarity between identity representations of each pair of face images. Please note that, the features from baselines (i.e. LBP, HoG, and MoCo) are also directly applied to perform verification for a fair comparison. As shown in Table 3, our identity representations can achieve 73.72% in accuracy on LFW, which outperforms the unsupervised state-of-the-art method [7].
LFW | CPLFW | |
Method | Accuracy(%) | Accuracy(%) |
Fully supervised | ||
VGG-Face [31] | 98.95 | 84.00 |
SphereFace [25] | 99.42 | 81.40 |
ArcFace [8] | 99.53 | 92.08 |
Unsupervised or hand-crafted features | ||
VGG [7] | 71.48 | - |
LBP | 56.90 | 51.50 |
HoG | 62.73 | 51.73 |
MoCo | 65.88 | 55.12 |
Ours | 73.72 | 58.52 |
3.5 Frontalization
Frontalization is the process of synthesizing the frontal facing view of a single facial image. In this work, there are two ways to obtain the neutral face in the frontal view: de-expression and re-identity. The de-expression operation removes the head motion and facial expressions from facial images and thus generates the neutral faces with frontal view. On the other hand, the re-identity operation recovers the neutral face by adding the identity to the mean face which is already in frontal view. As shown in Figure 5(a), the proposed method is able to synthesize the neutral faces from the facial images with various poses by the de-expression operation. The input images are from the LFW dataset [17] which are never seen during the training of our model. We also show a state-of-the-art approach in Figure 5(b) which additionally uses facial landmarks [14] for qualitative comparison.
We notice that the synthesized images from the proposed method are a little bit blurry, we hypothesize that it might be caused by the plenty blurry training images in the Voxceleb dataset. We believe that further improvements can be obtained by using other high-quality datasets.

3.6 Image-to-image Translation
The proposed model can naturally be used to perform image-to-image translation by transferring the facial motion of the source image into the target one. To this end, we simply calculate and then apply the backward flow field of the source image to warp the neutral face of the target image via the re-expression operation. As shown in Figure 6, our method can transfer the head pose and expression from the source to the target without noticeable artifacts. On the other hand, the results of X2Face method [42] reveal visible artifacts when the pose difference between source and target is large.

4 Conclusions
In this work, we propose novel cycle-consistency constraints for disentangling of identity and expression representations from a single facial image, that is, facial motion cycle-consistency and the identity cycle-consistency. The proposed model can be trained in an unsupervised manner by superimposing the proposed cycle-consistency constraints. We perform extensive qualitative and quantitative evaluations on multiple datasets to demonstrate the efficacy of our proposed method on learning disentangled facial representations. These representations contain rich and distinct information of identity and expression, and can be used to facilitate a variety of applications, such as facial expression recognition, head pose estimation, person recognition, frontalization, and the image-to-image translation.
Acknowledgement. This project is supported by MOST 108-2221-E-009-066-MY3, MOST 110-2636-E-009-001, and MOST 110-2634-F-009-018. We are grateful to the National Center for High-performance Computing for providing computing services and facilities.
References
- [1] Michael J Black and Paul Anandan. The robust estimation of multiple motions: Parametric and piecewise-smooth flow fields. Computer vision and image understanding, 63(1):75–104, 1996.
- [2] Volker Blanz, Thomas Vetter, et al. A morphable model for the synthesis of 3d faces. In ACM Transactions on Graphics (TOG), 1999.
- [3] Adrian Bulat and Georgios Tzimiropoulos. How far are we from solving the 2d & 3d face alignment problem?(and a dataset of 230,000 3d facial landmarks). In IEEE International Conference on Computer Vision (ICCV), 2017.
- [4] Baptiste Chu, Sami Romdhani, and Liming Chen. 3d-aided face recognition robust to expression and pose variations. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2014.
- [5] J. S. Chung, A. Nagrani, and A. Zisserman. Voxceleb2: Deep speaker recognition. In INTERSPEECH, 2018.
- [6] Navneet Dalal and Bill Triggs. Histograms of oriented gradients for human detection. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2005.
- [7] Samyak Datta, Gaurav Sharma, and CV Jawahar. Unsupervised learning of face representations. In IEEE International Conference on Automatic Face & Gesture Recognition (FG), 2018.
- [8] Jiankang Deng, Jia Guo, Niannan Xue, and Stefanos Zafeiriou. Arcface: Additive angular margin loss for deep face recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- [9] Paul Ekman and Wallace V Friesen. Facial action coding system: Investigator’s guide. Consulting Psychologists Press, 1978.
- [10] Corneliu Florea, Laura Florea, Mihai-Sorin Badea, Constantin Vertan, and Andrei Racoviteanu. Annealed label transfer for face expression recognition. In British Machine Vision Conference (BMVC), 2019.
- [11] Leon A Gatys, Alexander S Ecker, and Matthias Bethge. Image style transfer using convolutional neural networks. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
- [12] Ian J Goodfellow, Dumitru Erhan, Pierre Luc Carrier, Aaron Courville, Mehdi Mirza, Ben Hamner, Will Cukierski, Yichuan Tang, David Thaler, Dong-Hyun Lee, et al. Challenges in representation learning: A report on three machine learning contests. In International conference on neural information processing, 2013.
- [13] Michael E Hasselmo, Edmund T Rolls, and Gordon C Baylis. The role of expression and identity in the face-selective responses of neurons in the temporal visual cortex of the monkey. Behavioural Brain Research, 1989.
- [14] Tal Hassner, Shai Harel, Eran Paz, and Roee Enbar. Effective face frontalization in unconstrained images. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015.
- [15] James V Haxby, Elizabeth A Hoffman, and M Ida Gobbini. The distributed human neural system for face perception. Trends in Cognitive Sciences, 2000.
- [16] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
- [17] Gary B. Huang, Manu Ramesh, Tamara Berg, and Erik Learned-Miller. Labeled faces in the wild: A database for studying face recognition in unconstrained environments. Technical report, University of Massachusetts, Amherst, 2007.
- [18] Xun Huang and Serge Belongie. Arbitrary style transfer in real-time with adaptive instance normalization. In IEEE International Conference on Computer Vision (ICCV), 2017.
- [19] Junhwa Hur and Stefan Roth. Mirrorflow: Exploiting symmetries in joint optical flow and occlusion estimation. In IEEE International Conference on Computer Vision (ICCV), 2017.
- [20] Zi-Hang Jiang, Qianyi Wu, Keyu Chen, and Juyong Zhang. Disentangled representation learning for 3d face shape. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- [21] Justin Johnson, Alexandre Alahi, and Li Fei-Fei. Perceptual losses for real-time style transfer and super-resolution. In European Conference on Computer Vision (ECCV), pages 694–711. Springer, 2016.
- [22] Vahid Kazemi and Josephine Sullivan. One millisecond face alignment with an ensemble of regression trees. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2014.
- [23] Shan Li, Weihong Deng, and JunPing Du. Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- [24] Yong Li, Jiabei Zeng, Shiguang Shan, and Xilin Chen. Self-supervised representation learning from videos for facial action unit detection. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- [25] Weiyang Liu, Yandong Wen, Zhiding Yu, Ming Li, Bhiksha Raj, and Le Song. Sphereface: Deep hypersphere embedding for face recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- [26] Liupei Lu, Leili Tavabi, and Mohammad Soleymani. Self-supervised learning for facial action unit recognition through temporal consistency. In British Machine Vision Conference (BMVC), 2020.
- [27] Yongyi Lu, Yu-Wing Tai, and Chi-Keung Tang. Attribute-guided face generation using conditional cyclegan. In European Conference on Computer Vision (ECCV), 2018.
- [28] Kenji Mase. Recognition of facial expression from optical flow. IEICE TRANSACTIONS on Information and Systems, 1991.
- [29] A. Nagrani, J. S. Chung, and A. Zisserman. Voxceleb: a large-scale speaker identification dataset. In Annual Conference of the International Speech Communication Association (INTERSPEECH), 2017.
- [30] Timo Ojala, Matti Pietikainen, and Topi Maenpaa. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 24(7):971–987, 2002.
- [31] Omkar M Parkhi, Andrea Vedaldi, and Andrew Zisserman. Deep face recognition. 2015.
- [32] Albert Pumarola, Antonio Agudo, Aleix M Martinez, Alberto Sanfeliu, and Francesc Moreno-Noguer. Ganimation: Anatomically-aware facial animation from a single image. In European Conference on Computer Vision (ECCV), 2018.
- [33] Christos Sagonas, Georgios Tzimiropoulos, Stefanos Zafeiriou, and Maja Pantic. 300 faces in-the-wild challenge: The first facial landmark localization challenge. In IEEE International Conference on Computer Vision Workshops, 2013.
- [34] Wei Shen and Rujie Liu. Learning residual images for face attribute manipulation. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- [35] Zhixin Shu, Mihir Sahasrabudhe, Riza Alp Guler, Dimitris Samaras, Nikos Paragios, and Iasonas Kokkinos. Deforming autoencoders: Unsupervised disentangling of shape and appearance. In European Conference on Computer Vision (ECCV), 2018.
- [36] Zhixin Shu, Ersin Yumer, Sunil Hadap, Kalyan Sunkavalli, Eli Shechtman, and Dimitris Samaras. Neural face editing with intrinsic image disentangling. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- [37] Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. ArXiv:1409.1556, 2014.
- [38] Lawrence Sirovich and Michael Kirby. Low-dimensional procedure for the characterization of human faces. Journal of the Optical Society of America A, 1987.
- [39] Deqing Sun, Xiaodong Yang, Ming-Yu Liu, and Jan Kautz. Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
- [40] Luan Tran, Xi Yin, and Xiaoming Liu. Disentangled representation learning gan for pose-invariant face recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- [41] Olivia Wiles, A Koepke, and Andrew Zisserman. Self-supervised learning of a facial attribute embedding from video. In British Machine Vision Conference (BMVC), 2018.
- [42] Olivia Wiles, A Sophia Koepke, and Andrew Zisserman. X2face: A network for controlling face generation using images, audio, and pose codes. In European Conference on Computer Vision (ECCV), 2018.
- [43] Joel S Winston, RNA Henson, Miriam R Fine-Goulden, and Raymond J Dolan. fmri-adaptation reveals dissociable neural representations of identity and expression in face perception. Journal of Neurophysiology, 2004.
- [44] Xianglei Xing, Tian Han, Ruiqi Gao, Song-Chun Zhu, and Ying Nian Wu. Unsupervised disentangling of appearance and geometry by deformable generator network. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- [45] Huiyuan Yang, Umur Ciftci, and Lijun Yin. Facial expression recognition by de-expression residue learning. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
- [46] Tsun-Yi Yang, Yi-Ting Chen, Yen-Yu Lin, and Yung-Yu Chuang. Fsa-net: Learning fine-grained structure aggregation for head pose estimation from a single image. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2019.
- [47] Egor Zakharov, Aliaksandra Shysheya, Egor Burkov, and Victor Lempitsky. Few-shot adversarial learning of realistic neural talking head models. In IEEE International Conference on Computer Vision (ICCV), 2019.
- [48] Feifei Zhang, Tianzhu Zhang, Qirong Mao, and Changsheng Xu. Joint pose and expression modeling for facial expression recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
- [49] Yuting Zhang, Yijie Guo, Yixin Jin, Yijun Luo, Zhiyuan He, and Honglak Lee. Unsupervised discovery of object landmarks as structural representations. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018.
- [50] Shuwen Zhao, Haibin Cai, Honghai Liu, Jianhua Zhang, and Shengyong Chen. Feature selection mechanism in cnns for facial expression recognition. In British Machine Vision Conference (BMVC), 2018.
- [51] T. Zheng and W. Deng. Cross-pose lfw: A database for studying cross-pose face recognition in unconstrained environments. Technical Report 18-01, Beijing University of Posts and Telecommunications, February 2018.
- [52] Xiangyu Zhu, Zhen Lei, Xiaoming Liu, Hailin Shi, and Stan Z Li. Face alignment across large poses: A 3d solution. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.