Global-to-local Expression-aware Embeddings for Facial Action Unit Detection
Abstract
Expressions and facial action units (AUs) are two levels of facial behavior descriptors. Expression auxiliary information has been widely used to improve the AU detection performance. However, most existing expression representations utilized in AU detection works can only describe pre-determined discrete categories (e.g., Angry, Disgust, Happy, Sad, etc.) and cannot capture subtle expression transformations like AUs. In this paper, we propose a novel fine-grained Global Expression representation Encoder to capture subtle and continuous facial movements, to promote AU detection. To obtain such a global expression representation, we propose to train an expression embedding model on a large-scale expression dataset according to global expression similarity. Moreover, considering the local definition of AUs, it is essential to extract local AU features. Therefore, we design a Local AU Features Module to generate local facial features for each AU. Specifically, it consists of an AU feature map extractor and a corresponding AU mask extractor. First, the two extractors transform the global expression representation into AU feature maps and masks, respectively. Then, AU feature maps and their corresponding AU masks are multiplied to generate AU masked features focusing on local facial region. Finally, the AU masked features are fed into an AU classifier for judging the AU occurrence. Extensive experiment results demonstrate the superiority of our proposed method. Our method validly outperforms previous works and achieves state-of-the-art performances on widely-used face datasets, including BP4D, DISFA, and BP4D+.
Index Terms:
Facial action coding, facial action unit detection, facial expression recognition, expression-aware embedding, deep learning1 Introduction
Facial Affect Analysis (FAA) is an active research area in computer vision and affective computing communities. FAA includes two approaches, namely expression and Facial Action Units (AUs). As a universal non-verbal approach to carrying out human communication, facial expressions can be described via a combination of AUs [1]. Facial AUs, coded by the Facial Action Coding System (FACS) [1], include 32 atomic facial action descriptors based on anatomical facial muscle groups. Table I summarizes the information of the 15 AUs used in this paper, including their names and corresponding involved muscles. Each AU defines the movement of a specific face region. For example, AU1 (Inner Brow Raiser) is a descriptor focused on the medial part of the frontalis. AUs and expression are highly related to each other. Nearly any possible expression can be described as a specific combination of facial AUs. For instance, as shown in Figure 1, a happy expression can be achieved with the occurrence of both AU6 and AU12; a doubt expression is related to AU4, and a surprise expression is related to AU1 and AU26. From this perspective, AU features are often considered as a kind of facial expression representation. In fact, AUs play an important role in conveying human emotions and automatic expression analysis. AU detection has attracted much attention in recent years due to its wide applications, including emotion recognition [2], micro-expression detection [3], and mental health diagnosis [4].

AU detection is challenging due to the lack of annotated data and difficulty in capturing local features. Annotating AUs takes hours of tedious effort to manually label one hundred images, even for well-trained experts [5]. Therefore, it is time-consuming and laborious to obtain large scale well-annotated AUs data[6][7]. In terms the lack of labeled data, on the one hand, it often makes the training of deep networks suffer from data over-fitting. On the other hand, the AU datasets usually contain limited identities, dozens[8][9] or hundreds[10], which also leads to identity over-fitting[11][12][13]. Compared with AU datasets, there are amounts of accessible and easy-to-annotate expression data, like FEC[14] and AffectNet[15], which are beneficial for improving AU detection performance due to their close correlations [6][7][11]. However, simply regarding expressions as several rough discrete classes is sub-optimal for learning subtle expression transformations [16][17]. By contrast, in this paper we measure the subtle distinctions of expressions through similarity learning, leveraging continuous and compact expression features for AU representations.
As a facial AU is defined anatomically according to the movements of its corresponding facial muscles, it is inherently related to local facial regions, termed as AU’s locality. To this end, how to effectively extract local features from an input face image is a key yet unresolved challenge, because local facial deformations not only are subtle and transient [5][18] but also vary with individuals [11][12]. The straightforward approach is to divide the input image into several local parts in a pre-defined way, which are then fed into networks to capture local information, known as patch learning [19][20][21]. These methods pre-define AU local regions with prior knowledge, cropping a face image into several regular yet coarse-grained patches. Subsequently, the feature extractor is limited to the quality of cropped patches.
Another category of approaches is to utilize structured geometric information like landmarks to crop more fine-grained local partitions such as Regions Of Interest (ROIs) [22][23] or attention maps[24][25][26][27][28]. Some recent works[25][26][27] focus on learning attention maps based on facial landmarks in a supervised manner. The attention maps are usually initialized by pre-defined AU centers based on landmarks and treated as the ground truth during training. However, facial AUs occur in specific locations but are not limited to landmark points in most cases [29]. Hence, pre-defined AU regions or centers in a uniform way (either by landmarks or prior domain knowledge) are generally error-prone. For example, many prior AU patch generation procedures are fixed for various head poses, while landmark detection tends to deviate under large head rotations, leading to inaccurate AU patches or attention[21]. In sum, these approaches are sub-optimal in two aspects. First, AU centers based on domain knowledge and landmarks are limited to landmark positions and/or pre-defined rules[29][24]. Second, landmark detection errors often influence the final performance[29][24]. Therefore, capturing local features based on manually pre-defined rules with coarse guidances may limit the capability of the local feature extractor. To this end, the sensitive position of each AU in the face feature maps should depend on the AU detection task and the training dataset. In other words, the attention map learning progress should be data-driven.
Motivated by the above observations, we propose a novel AU detection framework, Global-to-Local Expression-aware Network(GTLE-Net), to address above mentioned challenges. Firstly, for the data and identity over-fiting challenge, we pre-train a global facial expression representation encoder(GEE) to extract identity-invariant global expression features on a large-scale facial expression dataset, FEC [14]. Instead of roughly classifying expressions into several discrete categories, we project expressions into a continuous representation space, which is more suitable and effective for AU detection in our experiments, since AUs are intrinsically subtle and continuous. Secondly, for the local feature capturing challenge, we design a local AU feature module (LAM) with two extractors to produce AU mask and feature map for each AU, respectively. Particularly, the AU mask extractor is learned without any intermediate supervision. Then, by multiplying the AU feature maps and the corresponding AU masks, we obtain the AU masked features, which filter out irrelevant fields and retain informative ones. In this way, our framework can not only make full use of large scale well-annotated expression data but also learn attention maps adaptively without extra supervision, to promote AU detection .
To sum up, the main contributions of this paper can be summarized below:
-
(1)
A global expression feature encoder(GEE) is pre-trained with structured triplet data. This step is proved to be essential for the improvement of AU detection.
-
(2)
An effective attention learning method by a local AU features module (LAM) is proposed to capture AU-specific representations, which is crucial to obtain more informative local features and boost up AU detection performance. Visual analytics prove that LAM produces highly AU-related local information adaptively.
-
(3)
Through extensive experiments, we validate the effectiveness and accuracy of the proposed method on three widely-used benchmark datasets, outperforming the state-of-the-art significantly.
The remainder of this paper is organized as follows. Related works are presented in Section 2. The details of our method are described in Section 3. Comprehensive experiment results are reported in Section 4. Section 5 provides a series of visual analyses. Discussion and concluding remarks are provided in Section 6. Our code will be released if accepted.
AU Name | Description | Involved Facial Muscle(s) |
---|---|---|
AU1 | Inner Brow Raiser | Frontalis, pars medialis |
AU2 | Outer Brow Raiser | Frontalis, pars lateralis |
AU4 | Brow Lowerer | Depressor Glabellae |
AU6 | Cheek Raiser | Orbicularis oculi |
AU7 | Lid Tightener | Orbicularis oculi |
AU9 | Nose Wrinkler | Levator labii superioris alaquae nasi |
AU10 | Upper Lip Raiser | Levator Labii Superioris |
AU12 | Lip Corner Puller | Zygomatic Major |
AU14 | Dimpler | Buccinator |
AU15 | Lip Corner Depressor | Depressor anguli oris |
AU17 | Chin Raiser | Mentalis |
AU23 | Lip Tightener | Orbicularis oris |
AU24 | Lip Pressor | Orbicularis oris |
AU25 | Lips part | Depressor labii inferioris |
AU26 | Jaw Drop | Masseter |
2 Related Work
In recent years, researchers have developed many approaches for AU detection including multi-task and attention mechanisms, introducing some auxiliary information like landmarks, text description, and expressions, etc. We describe previous related works below.
2.1 AU Detection with Auxiliary Information
Due to the high labor cost of AU annotations, the scale and subject variations of AU detection datasets are usually limited. As a result, previous AU detection methods resort to various kinds of auxiliary information to improve the generalization performance. Facial landmarks are the most widely used pre-trained features for AU detection. JPML [30] employs facial landmark features to crop local image patches for different AUs. EAC-Net [25] constructs local regions of interest and spatial attention maps from the facial landmarks. LP-Net [12] trains an individual-specific shape regularization network from the detected facial landmarks. JÂA-Net [26] jointly performs AU detection and facial landmark detection from the data annotated with both labels.
Other types of auxiliary information have also been explored. Zhao et al. [31] pre-train a weakly supervised embedding from a large number of web images. Cui et al.[16] summarize the prior probabilities of AU occurrences as generic knowledge. Emotions and AUs are jointly optimized under the prior probabilities. Recently, SEV-Net [28] leverages textual descriptions of AU occurrences by employing a pre-trained word embedding to obtain auxiliary textual features. To enhance the generalization of our model, our method introduces a pre-trained expression embedding as the auxiliary information.
2.2 AU Feature Learning
Due to the local definitions of AUs, it is essential to extract local AU features. As such, some researchers proposed to obtain local information through patch learning. For instance, Zhong et al. [19] and Liu et al. [20] preprocess an input image into uniform patches before encoding to analyze facial expressions. Taking the head pose into consideration, Onal et al.[21, 32] crop AU-specific local facial patches containing information for specific AU recognition after registering the 3D head pose to reduce changes from head movements. Besides, it is a common practice to use attention mechanisms to highlight the features at the facial AU-based positions. Facial landmarks with sparse facial geometric features show an advantage in operating as a supervised attention prior. EAC-Net [25] creates fixed attention maps related to the correlations between AUs and landmarks. JÂA-Net [26] jointly performs AU detection and facial landmark detection, and the predicted landmarks are used to compute the attention map for each AU. Jacob et al. [27] propose a multi-task method that combines the tasks of AU detection and landmark-based attention map prediction. ARL et al. [24] proposed a channel-wise and spatial attention learning for each AU. And, a pixel-level relation is learned by CRF to refine the spatial attention. Except for the facial landmarks, SEV-Net [28] utilizes the textual descriptions of local details to generate a regional attention map. In this way, it highlights the local parts of global features. However, it requires extra annotations for descriptions. Our work proposes a pixel-wise self-attention map that is data-driven without supervision. Our experiments prove that our attention map is superior to the attention maps produced by prior knowledge.
2.3 Expression Representations
Expressions are essential auxiliary information for AU detection. Some prior works leverage amounts of accessible expression data to enhance AU detection[11][7][31][16]. Recently, Chang et al.[33] utilize a large amount of unlabeled images to train a representation encoder to extract local representations and project them to a low-dimensional latent space, then improve network performance through contrastive learning. Many methods map face images into a low-dimensional manifold for subject-independent expression representations. Early works [15][34][35] train the embeddings for discrete emotion classification tasks but neglect the facial expression variations within each class. The 3D Morphable Model (3DMM) [36][37] has been proposed to fit identities and expression parameters from a single face image. Expressions are represented as the coefficients of predefined blendshapes in 3DMM. The estimated expression coefficients are then used for talking head synthesis [38][39], expression transfer [40][41], and face manipulation [42]. However, the estimated expression coefficients have weaknesses in representing fine-grained expressions. To solve the problem, Vemulapalli and Agarwala [14] proposed a compact embedding for complicated and subtle facial expressions, where facial expression similarity is defined through triplet annotations. Zhang et al. [43] proposed a Deviation Learning Network (DLN) to remove the identity information from continuous expression embeddings, and thus achieve more compact and smooth representations.
3 Proposed Method
In this section, we first briefly introduce the problem definition and then describe our proposed GTLE-Net framework.

3.1 Problem Definition
The task of AU detection is to predict the AU occurrence probabilities given an input face image with a resolution of . As described in section 1, expressions and AUs are two means of description from global and local levels respectively. They are highly related and thus can be utilized to promote each other. Moreover, facial expression data are easier to obtain and annotate, while AUs are more subtle and difficult to annotate. Intuitively, facial AU information can be extracted from facial expression representations when expression representation perception is adequately fine-grained. Considering the above factors, we propose a global expression encoder(GEE) and pre-train it on a large-scale facial expression dataset, aiming to obtain a powerful and robust expression representation. Furthermore, for the purpose of capturing AU local features, we propose a local AU features module (LAM) constructed with two extractors, namely, the AU mask extractor () and the AU feature map extractor (). The framework is illustrated in Fig 2. In the following, we discuss each main module in detail.
3.2 Global Expression Encoder
To enhance the generalization of our model, we introduce a compact and continuous expression embedding [14][43] as the prior auxiliary information. Specifically, GEE was pre-trained on the expression similarity task (See Fig. 3) using the FEC dataset [14]. The FEC dataset contains numerous facial image triplets along with the annotations indicating which one has the most different expression in each triplet. We pre-trained using the triplet loss to constrain that more similar expressions have a smaller embedding distance. In each triplet, we call the image with the most different expression as the Negative () and the other two images the Anchor () and Positive (), respectively. For a triplet (A, P, N), our outputs three expression embeddings , and . The pre-trained training process utilizes the triplet loss as follows:
(1) | ||||
Due to the large number of triplets, the pre-trained feature encoder can produce a compact and identity-invariant expression representation, which is beneficial for the downstream expression-related tasks like AU detection.
We choose the structure of InceptionResnetV1 [44] as our backbone of , replacing the last classifier layer with a linear layer followed by normalization. After the dimensionality reduction, it outputs a 16-dimensional vector as the expression embedding. We first pre-train on the expression similarity task and then choose the shallow layers (up to the layer with feature maps in a resolution of ) to extract the global expression representation from the input face image , which is illustrated as:
(2) |
Then, our whole framework including is fine-tuned for AU detection.
Our intuition is twofold. First, initialized from expression embedding learning could instruct the AU detection framework to capture more effective expression features. Second, it obtains a good initialization at the beginning of training, which helps to relieve the problem of over-fitting caused by the limited data and identities of AU datasets.

3.3 Local AU Features Module
The expression features extracted by the pre-trained describe facial movements globally. However, this global representation may not be suitable directly for the subsequent AU detection task, since AUs are defined locally and focus on subtle local motions. Therefore, it is more appropriate to extract local fine-grained features for each AU individually. For this purpose, we design the with two parallel extractors (i.e., a AU mask extractor() and an AU feature map extractor()) to transform the global expression features into local AU-related representations.
Specifically, we first extract from the global expression features to produce AU features with our designed AU feature map extractor . Each AU feature is related to a specific AU. However, we found that the AU features generated by the may contain information from unrelated AUs, so we introduce an AU mask extractor() to find the region of interest for each AU and also mask out irrelevant content.
AU Feature Map Extractor. The AU feature map extractor aims at generating AU-related features. Given the global expression feature , our AU feature map extractor () performs as follows:
(3) |
where is the -th AU feature map that is related to the -th AU, and is the total number of all AUs. represents the content that is not related to any AUs (e.g., background).
Considering that there are many AUs on the face, if we design a extracting branch for each AU individually, there will be a lot of parameters. Therefore, we design the AU feature map extractor to be a shared backbone with multiple prediction heads. The shared backbone is formed by stacking multiple residual blocks [45] and upsampling layers. The upsampling design can enlarge some local facial details that benefit the subsequent AU detection and reconstruction tasks. Each of the prediction heads consists of a convolutional layer and an activation layer, and it is responsible for predicting the facial content related to a specific AU.
AU Mask Extractor. The AU mask extractor is responsible for generating AU masks maps from the global expression features . Each mask represents the region of interest of each AU. Given the global features , the mask maps are predicted by:
(4) |
where is the -th facial mask for the -th AU, and is for the background.
To ensure that the values of the predicted facial masks lie in a fixed range of 0 to 1 (0 denotes “no attention” and 1 denotes “the strongest attention”), an activation function is added to the last layer of . Intuitively, Sigmoid and Softmax are both suitable for this purpose. However, Softmax tend to get sparse attention maps and exclude regions related to the target AU. This may result in the downstream recognition task being unable to learn sufficient information. Hence, we choose Sigmoid to ativate the outputs of
AU Masked Features. Once obtaining the AU specific feature maps and masks, the final AU masked features are generated by multiplying them:
(5) |
where () is the computed masked AU feature for the -th AU that is then used for the subsequent AU detection task. While is the masked background content which is used for the subsequent face image reconstruction task.
To ensure these AU feature maps contain all facial features of the input image, we introduce a reconstruction task. Specifically, we design an image reconstruction head (), which is fed with all masked AU features, and we expect the reconstruction result to be the same as the input facial image.
(6) |
where is the reconstruction result.
In this way, the not only realizes the generation of AU local/masked features but also retains the complete information of the input face image. Finally, the generated AU masked features are sent to an AU Classifier (described below) for AU recognition.
3.4 AU Classifier
Since the outputs of the are a series of RGB feature maps with a high resolution (), and the number of the feature maps equals the number of AUs in the dataset, the following AU classifier module, denoted as , is designed to be a multi-layer CNN architecture, which contains 6 convolution blocks and a global average pooling layer. Each of the first five convolution layers is followed by a batch-norm and an ELU activation function, and their output channels are set to 16, 32, 64, 128, and 256, respectively. To make sure the sub-module has a wide receptive field, the five CNNs’ kernel size, stride, and padding are set to 7, 2, and 3, respectively. The global average pooling layer is set after the fifth convolution block, aiming to aggregate the feature maps into AU Embeddings. In this way, more information closely related to the specific AU will be extracted from the input AU masked feature map. After that, the AU embeddings are fed into the last convolution layers, where all of the output channel, kernel size, stride, and padding are set to 1. This procedure can be formulated as the following function:
(7) |
Here indicates the output of the last convolution layer of the -th AU. To obtain the final prediction of the AU’s probability, we use a Sigmoid activation on the and obtain the predicted probability as follows:
(8) |
3.5 Loss Functions
As illustrated in Figure 2, our proposed framework refers to two steps. The pretraining procedure is expression similarity learning, while the AU detection part contains two tasks, namely, AU classification and facial image reconstruction. The pretraining loss is a triplet loss, referring to Equation (1), and the AU detection training loss contains two parts as follows:
(9) |
where is the AU classification loss and is the facial image reconstruction loss, is a hyperparameter for adjusting the loss ratio. Our main goal is to recognize the AU, the image reconstruction is merely an auxiliary task, so the is set to a quite low vale of 0.001 in our experiments.
AU detection can be conceptually regarded as a multi-label binary classification problem, so the AU detection loss is defined as:
(10) |
where is the number of AU classes, is the ground truth, and is the predicted probability of the -th AU. As for the image reconstruction loss , we use a pixel-level MSE loss between the input image and the reconstructed one as follows:
(11) |
where is the pixels number of input image and equals to . Besides, and are the ground-truth and reconstructed value of the -th pixel of the input image, respectively.
4 Experimental Results
In this section we report extensive experiments to validate the effectiveness of our method. These experiments are conducted on three public benchmarks commonly-used for AU detection, all of which contain face expression images with manually-annotated AU labels.
We first introduce the used three datasets in detail and then describe the implementation details of the training and validation processes. Furthermore, we compare our method with the start-of-the-art methods to demonstrate the superiority of our approach. For fair comparisons, our evaluation follows the prior works [25, 26], using three-fold cross-validation and also their specific division of training and validation data (see below). The reported results are the averages from three-fold experiments. To further obtain insight into our method, we designed ablation experiments to study the impact of the built sub-modules. Finally, we visualize several samples to further validate our method.
4.1 Datasets
As mentioned above, our evaluation was performed on three widely-used AU detection datasets: BP4D [8], DISFA [9], and BP4D+ [10]. All the above datasets were annotated with AU labels by certified experts for each frame.
BP4D consists of 328 video clips from 41 participants(23 females and 18 males). Each video frame was annotated with the occurrence or absence of 12 AUs(1, 2, 4, 6, 7, 10, 12, 14, 15, 17, 23, and 24. Their definition can be seen in Table I). In total, about 140,000 frames were annotated frame-by-frame by a group of FACS experts. Following the same subject division [25, 26], we performed three-fold, subject-exclusive, cross-validations for all experiments.
Method | AU1 | AU2 | AU4 | AU6 | AU7 | AU10 | AU12 | AU14 | AU15 | AU17 | AU23 | AU24 | Avg. |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
LSVM | 23.2 | 22.8 | 23.1 | 27.2 | 47.1 | 77.2 | 63.7 | 64.3 | 18.4 | 33.0 | 19.4 | 20.7 | 35.3 |
JPML | 32.6 | 25.6 | 37.4 | 42.3 | 50.5 | 72.2 | 74.1 | 65.7 | 38.1 | 40.0 | 30.4 | 42.3 | 45.9 |
DRML | 36.4 | 41.8 | 43.0 | 55.0 | 67.0 | 66.3 | 65.8 | 54.1 | 33.2 | 48.0 | 31.7 | 30.0 | 48.3 |
EAC-Net | 39.0 | 35.2 | 48.6 | 76.1 | 72.9 | 81.9 | 86.2 | 58.8 | 37.5 | 59.1 | 35.9 | 35.8 | 55.9 |
DSIN | 51.7 | 40.4 | 56.0 | 76.1 | 73.5 | 79.9 | 85.4 | 62.7 | 37.3 | 62.9 | 38.8 | 41.6 | 58.9 |
ARL | 45.8 | 39.8 | 55.1 | 75.7 | 77.2 | 82.3 | 86.6 | 58.8 | 47.6 | 62.1 | 47.4 | 55.4 | 61.1 |
LP-Net | 43.4 | 38.0 | 54.2 | 77.1 | 76.7 | 83.8 | 87.2 | 63.3 | 45.3 | 60.5 | 48.1 | 54.2 | 61.0 |
AU-GCN | 46.8 | 38.5 | 60.1 | [80.1] | 79.5 | 84.8 | 88.0 | 67.3 | 52.0 | 63.2 | 40.9 | 52.8 | 62.8 |
SRERL | 46.9 | 45.3 | 55.6 | 77.1 | 78.4 | 83.5 | 87.6 | 63.9 | 52.2 | 63.9 | 47.1 | 53.3 | 62.9 |
AU-RCNN | 50.2 | 43.7 | 57 | 78.5 | 78.5 | 82.6 | 87 | 67.7 | 49.1 | 62.4 | 50.4 | 49.3 | 63.0 |
UGN-B | 54.2 | 46.4 | 56.8 | 76.2 | 76.7 | 82.4 | 86.1 | 64.7 | 51.2 | 63.1 | 48.5 | 53.6 | 63.3 |
JÂA-Net | 53.8 | 47.8 | 58.2 | 78.5 | 75.8 | 82.7 | 88.2 | 63.7 | 43.3 | 61.8 | 45.6 | 49.9 | 62.4 |
SEV-Net | 58.2 | 50.4 | 58.3 | 81.9 | 73.9 | 87.8 | 87.5 | 61.6 | [52.6] | 62.2 | 44.6 | 47.6 | 63.9 |
HMP-PS | 53.1 | 46.1 | 56.0 | 76.5 | 76.9 | 82.1 | 86.4 | 64.8 | 51.5 | 63.0 | [49.9] | 54.5 | 63.4 |
MHSA-FFN | 51.7 | [49.3] | [61.0] | 77.8 | 79.5 | 82.9 | 86.3 | 67.6 | 51.9 | 63.0 | 43.7 | [56.3] | 64.2 |
CISNet | 54.8 | 48.3 | 57.2 | 76.2 | 76.5 | 85.2 | 87.2 | 66.2 | 50.9 | [65.0] | 47.7 | 56.5 | 64.3 |
D-PAttNe | 50.7 | 42.5 | 59.0 | 79.4 | 79.0 | 85.0 | [89.3] | 67.6 | 51.6 | 65.3 | 49.6 | 54.5 | 64.7 |
CaFNet | [55.1] | 49.3 | 57.7 | 78.3 | 78.6 | [85.1] | 86.2 | [67.4] | 52.0 | 64.4 | 48.3 | 56.2 | [64.9] |
GTLE-Net | 58.2 | 48.7 | 61.5 | 78.7 | [79.2] | 84.2 | 89.8 | 66.3 | 56.7 | 64.8 | 53.5 | 53.6 | 66.3 |
DISFA contains 27 video clips from 27 participants(12 females and 15 males). Each video records the facial activity of a participant when watching a 4-minutes video clip. Each clip consists of 4,845 frames evenly. Altogether, around 130,000 frames were annotated with intensities from 0 to 5. Following previous works [27][25][26][28], we regard intensity labels as absence and others as presence. The evaluation was performed on 8 AUs(1, 2, 4, 6, 9, 12, 25, and 26. Their definition can be seen in Table I). We also performed the same three-fold, subject-exclusive, cross validations on 8 AUs as in the previous works [25][26].
BP4D+ contains 1400 video clips from 140 participants(82 females and 58 males). Compared to BP4D, it has larger identity variations and video numbers. About 198,000 frames were annotated with the same 12 AUs in BP4D. We performed three-fold, subject-exclusive, cross-validations to compare our method with the state-of-the-art. Besides, to evaluate the generalization performance, our method was trained on BP4D and then evaluated on BP4D+, following the previous works [24][26].
To sum up, the order from largest to smallest of the three datasets is BP4D+, BP4D and DISFA. BP4D and BP4D+ were recorded in similar shooting environments and lighting conditions. Both of them share the same 12 AU classes. In comparison, DISFA is the smallest dataset containing only 8 AU classes and was captured in quite different lighting conditions.
Method | AU1 | AU2 | AU4 | AU6 | AU9 | AU12 | AU25 | AU26 | Avg. |
---|---|---|---|---|---|---|---|---|---|
LSVM | 10.8 | 10.0 | 21.8 | 15.7 | 11.5 | 70.4 | 12.0 | 22.1 | 21.8 |
DRML | 17.3 | 17.7 | 37.4 | 29.0 | 10.7 | 37.7 | 38.5 | 20.1 | 26.7 |
APL | 11.4 | 12.0 | 30.1 | 12.4 | 10.1 | 65.9 | 21.4 | 26.9 | 23.8 |
EAC-Net | 41.5 | 26.4 | 66.4 | 50.7 | 8.5 | 89.3 | 88.9 | 15.6 | 48.5 |
DSIN | 42.4 | 39.0 | 68.4 | 28.6 | 46.8 | 70.8 | 90.4 | 42.2 | 53.6 |
ARL | 43.9 | 42.1 | 63.6 | 41.8 | 40.0 | 76.2 | [95.2] | 66.8 | 58.7 |
LP-Net | 29.9 | 24.7 | 72.7 | 46.8 | 49.6 | 72.9 | 93.8 | 56.0 | 56.9 |
AU-GCN | 32.3 | 19.5 | 55.7 | 57.9 | 61.4 | 62.7 | 90.9 | 60.0 | 55.0 |
AU-RCNN | 32.1 | 25.9 | 59.8 | 55.4 | 39.8 | 67.7 | 77.4 | 52.6 | 51.3 |
SRERL | 45.7 | 47.8 | 56.9 | 47.1 | 45.6 | 73.5 | 84.3 | 43.6 | 55.9 |
UGN-B | 43.3 | 48.1 | 63.4 | 49.5 | 48.2 | 72.9 | 90.8 | 59.0 | 60.0 |
JÂA-Net | [62.4] | [60.7] | 67.1 | 41.1 | 45.1 | 73.5 | 90.9 | [67.4] | 63.5 |
MHSA-FFN | 46.1 | 48.6 | 72.8 | 56.7 | 50.0 | 72.1 | 90.8 | 55.4 | 61.5 |
SEV-Net | 55.3 | 53.1 | 61.5 | [53.6] | 38.2 | 71.6 | 95.7 | 41.5 | 58.8 |
HMP-PS | 38.0 | 45.9 | 65.2 | 50.9 | 50.8 | 76.0 | 93.3 | 67.6 | 61.0 |
CISNet | 48.8 | 50.4 | [78.9] | 51.9 | 47.1 | [80.1] | [95.4] | 65.0 | 64.7 |
CaFNet | 45.6 | 55.7 | 80.2 | 51.0 | 54.7 | 79.0 | 95.2 | 65.3 | [65.8] |
GTLE-Net | 63.5 | 62.5 | 69.8 | 48.6 | [53.9] | 75.6 | 94.7 | 64.4 | 66.6 |
Method | AU1 | AU2 | AU4 | AU6 | AU7 | AU10 | AU12 | AU14 | AU15 | AU17 | AU23 | AU24 | Avg. |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
FACS3D-Net | 43.0 | 38.1 | 49.9 | 82.3 | 85.1 | 87.2 | 87.5 | 66.0 | 48.4 | 47.4 | 50.0 | [31.9] | 59.7 |
ML-GCN | 40.2 | 36.9 | 32.5 | 84.8 | 88.9 | 89.6 | [89.3] | 81.2 | [53.3] | 43.1 | 55.9 | 28.3 | 60.3 |
MS-CAM | 38.3 | 37.6 | 25.0 | 85.0 | 90.9 | 90.9 | 89.0 | 81.5 | 60.9 | 40.6 | 58.2 | 28.0 | 60.5 |
SEV-Net | [47.9] | 40.8 | 31.2 | 86.9 | 87.5 | 89.7 | 88.9 | 82.6 | 39.9 | 55.6 | [59.4] | 27.1 | [61.5] |
GTLE-Net(Ours) | 51.5 | [46.6] | [43.5] | [86.8] | [89.6] | [91.0] | 89.8 | [82.3] | 46.8 | [49.3] | 60.9 | 50.9 | 65.7 |
Method | AU1 | AU2 | AU4 | AU6 | AU7 | AU10 | AU12 | AU14 | AU15 | AU17 | AU23 | AU24 | Avg. |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
EAC-Net | 38.0 | [37.5] | [32.6] | 82.0 | 83.4 | 87.1 | 85.1 | 62.1 | 44.5 | 43.6 | 45.0 | 32.8 | 56.1 |
ARL | 29.9 | 33.1 | 27.1 | 81.5 | 83.0 | 84.8 | 86.2 | 59.7 | [44.6] | 43.7 | 48.8 | 32.3 | 54.6 |
JÂA-Net | 39.7 | 35.6 | 30.7 | [82.4] | [84.7] | [88.8] | [87.0] | [62.2] | 38.9 | [46.4] | [48.9] | 36.6 | 56.8 |
GTLE-Net(Ours) | [39.5] | 37.7 | 44.0 | 84.5 | 84.9 | 89.6 | 87.7 | 72.3 | 44.9 | [45.3] | 52.6 | [33.9] | 59.7 |
Method | AU1 | AU2 | AU4 | AU6 | AU7 | AU10 | AU12 | AU14 | AU15 | AU17 | AU23 | AU24 | Avg. |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
w/o pretrain | 34.9 | 24.2 | 36.7 | 76.1 | 71.4 | 81.0 | 86.3 | 58.7 | 25.9 | 51.9 | 27.4 | 31.6 | 50.5 |
w. GEE Emotion | 48.1 | 41.6 | 53.9 | 78.6 | 77.9 | 83.2 | 88.8 | 60.5 | 39.3 | 65.5 | 38.0 | 47.6 | 60.3 |
w. GEE ImageNet | 49.6 | 39.7 | 60.5 | 77.3 | 78.7 | 83.1 | 88.6 | [64.6] | 48.2 | 63.2 | 47.5 | 48.0 | 62.4 |
w. GEE fixed | 52.3 | 45.8 | 59.8 | [79.4] | 78.1 | 84.8 | 89.0 | 64.0 | 46.5 | 64.0 | 38.6 | 49.8 | 63.5 |
w/o & | [55.4] | 65.1 | 73.9 | 49.0 | 46.9 | 77.1 | 92.9 | 54.1 | 48.1 | 64.7 | [52.5] | 47.5 | 63.3 |
w/o | 53.3 | 44.2 | 58.7 | 78.6 | 76.9 | 83.9 | [89.8] | 62.2 | 48.9 | 65.1 | [52.5] | 51.8 | 63.8 |
w/o & w. CNN | 51.8 | 43.0 | 59.1 | 78.3 | 78.1 | 84.2 | 88.7 | 63.1 | 50.2 | 64.0 | 49.2 | 55.1 | 63.7 |
GTLE-Net(Full) | 58.2 | [48.7] | [61.5] | 78.7 | [79.2] | 84.2 | [89.8] | 66.3 | 56.7 | 64.8 | 53.5 | [53.6] | 66.3 |
4.2 Implementation Details
Face images were first aligned and cropped according to landmark detection results. Before feeding to our networks, the input images were resized to . The resolution of feature maps was reduced to after passing through the .
To enhance the feature extraction capability of low-level convolution layers, the was initialized with the corresponding network parameters pretrained on the dataset of ImageNet or FEC, and the rest of our networks were set to Kaiming initialization [46]. Except for the AU Mask Extractor(), the whole model parameters were updated by employing an Adam optimizer with hyper-parameters and weight decay .
The networks were trained for 20 epochs with an initial learning rate of on each dataset fold. The mini-batch size was set to 10. We used a warm-up stage with steps and then decayed the learning rate exponentially with an exponent . The training and inference pipelines were implemented in Pytorch on 4 GeForce GTX 2080ti GPU.
4.3 Comparisons with the State-of-the-Art
We compared our method with many state-of-the-art AU detection methods, including LSVM [47], JPML [30], DRML [48], APL [49], EAC-Net [25], DSIN [50], ARL [24], SRERL [51], AU-GCN[23],AU-RCNN[29], ML-GCN [52], MS-CAM [53], LP-Net [12], FACS3D-Net [54], UGN-B [55], JÂA-Net [26], MHSA-FFN [27], SEV-Net [28], HMP-PS [56], D-PAttNe[21], CISNet[13] and CaFNet[57].
Specifically, while EAC-Net, ARL, JÂA-Net, and SEV-Net were evaluated on all three datasets, LSVM, JPML, DRML, APL, DSIN, SRERL, AU-GCN, AU-RCNN, LP-Net, UGN-B, MHSA-FFN, HMP-PS, CISNet and CaFNet were evaluated on both BP4D and DISFA. Particularly, D-PAttNe was only evaluated on BP4D. FACS3D-Net, ML-GCN and MS-CAM were evaluated merely on BP4D+. Since BP4D+ is released later, it is not used by all the methods.
Following the experimental settings of most previous works, we also employed the F1 score [58] on the frame level as the evaluation metric. F1-score is a harmonic mean of precision(P) and recall(R), which is defined as . Since F1-score takes both false positives and false negatives into account, it reflects a more rational measurement of the model’s performance than accuracy, especially on datasets with uneven class distributions like BP4D, DISFA and BP4D+.
Evaluation on BP4D. The results on BP4D are shown in Table II. The average F1 score of our method is 66.3%, which significantly outperforms all the other comparison methods, improving about 1.4% over the second highest one, CaFNet. In terms of the performances on single AUs, our method achieves the highest F1 scores among all the methods on 5 AUs(i.e., AU1, AU4, AU12, AU15, and AU23) and second highest on 1 AU(AU7) of the 12 evaluated AUs. Besides, our method shows over 3.5% superiority on AU15 and AU23 over the second highest ones.
In terms of AU feature learning, our method outperforms JMPL, DRML, and D-PAttNe about 20.4%, 18% and 1.6%, demonstrating that extracting local features by attention is more effective and our method is superior in AU feature extraction than patch learning.
Compared with EAC-Net, JAA-Net and MHSA-FFN, which utilize attention mechanisms with the supervisions of pre-defined AU centers, our method achieves improvements of 10.4%, 5.2% and 3.9% over them, respectively. This validates the superiority of our non-predetermined AU attention learning approach.
Evaluation on DISFA. Table III shows the evaluation results on DISFA. As shown in this table, our method achieves the average F1 score of 66.6%, outperforming all the compared state-of-the-art methods.
Compared to the most recent works SEV-Net, HMP-PS, MHSA-FFN, CISNet and CaFNet, our method achieves an improvement of 7.8%, 5.6%, 5.1%, 1.9% and 0.8% over them, respectively. Unlike EAC-Net and JÂA-Net which are supervised by both AU and landmark labels, our method only utilizes AU labels, but still obtains significant improvements of 18.1% and 3.1% over EAC-Net and JÂA-Net, respectively, demonstrating the effectiveness of our method in AU feature and attention learning. Note that our method does not explicitly model AU relationships, but significantly exceeds all relation-based methods including JMPL, DSIN, ARL, AU-GCN and AU-RCNN.


Evaluation on BP4D+. Table IV shows the evaluation results of the three-fold cross-validation on BP4D+. In particular, the results of SEV-Net, ML-DCN and MS-CAM are reported in[28] while the results of EAC-Net are reported in [26]. Our method obtains the highest or the second highest results on 11 of the 12 evaluated AUs. It achieves an average F1 score of 65.7% and outperforms all the state-of-the-art methods, exceeding the second highest method(SEV-Net) over 4.2%.
To further evaluate the generalization performance of our method on large-scale test identities, we trained our method on BP4D and evaluated it on the full BP4D+ of 140 subjects. The results are shown in Table V. We use the reported results in the work [26] for comparison. Again, our method outperforms all the compared methods under the large-scale cross-dataset evaluation. It indicates that our method can extract identity-independent information and generalize well to new identities.
4.4 Ablation Study
In this section, we conduct ablation studies to evaluate the effectiveness of each key sub-module in our framework, including: 1) the prior knowledge of expression similarity learning, and 2) the local AU features module. In particular, all the ablation experiments are conducted on the BP4D dataset due to space limitations. The experimental results are reported in Table VI.
Prior Knowledge. Our framework proposes a vision feature encoder that is pretrained on the FEC dataset via expression similarity learning, which refers to our prior knowledge. To test the efficiency of prior knowledge, we design ablation experiments with different approaches to initialize in the training of GTLE-Net, including:
-
a)
w/o pretrain: is randomly initialized;
-
b)
w.GEE Emotion: is initialized via a pretrained classifier of seven discrete facial expression emotion (e.g. neutral, happy, sad, surprise, fear, disgust, angry) categories trained on the AffectNet dataset;
-
c)
w. GEE ImageNet: is initialized via a pretrained image object classifier built on imageNet [59];
-
d)
GTLE-Net(Full): our proposed method, is initialized via the facial expression similarity learning;
-
e)
w. GEE fixed: is parameterized by the facial expression similarity approach but not updated in the training of GTLE-Net.
Firstly, in comparison with w/o pretrain, w.GEE Emotion, w. GEE ImageNet and w. GEE fixed achieve improvements of 9.8%(from 50.5% to 60.3%), 11.9%(from 50.5% to 62.4%) and 15.8%(from 50.5% to 66.3%), respectively. This indicates that the prior knowledge is necessary and the pretraining progress can bring significant improvement for AU detection.
Secondly, compared with the w. GEE Emotion and w. GEE ImageNet, GTLE-Net obtains the best performance(66.3%), exceeding the other two pretraining methods about 6% and 3.9%. This proves the effectiveness and superiority of the prior knowledge captured by our facial expression similarity learning.
Lastly, compared with GTLE-Net, w. GEE fixed obtains a lower average F1 of 63.5%. It indicates that finetuning the global expression feature encoder is beneficial for extracting features which are more adaptive to the following LAM module. Thus finetuning operation is essential and efficient to improve the performance of our framework.
Facial Components. To validate the effectiveness of the representations extracted by our local AU features module (LAM), we conduct three ablation experiments. Concretely, based on our full framework of GTLE-Net, we remove or replace some specific sub-modules and retrain the whole network, including:
-
f)
w/o &: removing both the and the ;
-
g)
w/o : only removing ;
-
h)
w/o & w. CNN: removing the and replacing the with a CNN block in the meanwhile. Particularly, the CNN block has similar number of parameters(around 16.6M) with our feature encoder().
Note that the inclusion of alone(w/o ) brings an improvement of 0.5%(from 63.3% to 63.8%) in comparision to w/o &. When the is replaced by the CNN module(w/o & w. CNN ), the performance declines by 0.1%(from 63.8% to 63.7%). Moreover, the F1-Score improves by 2.5% when we further add the module(GTLE-Net). These study results validate the effectiveness of our AU mask and local feature learning.

4.5 Visual Analysis
We also visualize the feature and mask results of our framework to further validate its effectiveness.
Mask Attention Learning. Due to its anatomical definition, each AU involves a group of muscle motions and is related to a specific region on the face. For instance, AU1 is inner brow raising and AU9 is nose wrinkle. Theoretically our framework should learn a regional mask activation map for each AU, corresponding to its muscle actions.
Figure 5 visualizes the attention mask maps of several examples from DISFA, BP4D and BP4D+, learned by our network. As shown in this figure, we can see that our method successfully captures a discriminative attention mask for these AUs. Taking AU4 (brow low) and AU17 (chin raise) as examples, the learned spatial masks are around the brows and chin, respectively, which demonstrates that our proposed AU mask extractor is effective for AU-related local regional attention learning.
From Figure 4, we can see that each AU’s mask is closely related to its muscle region, even for different expressions such as happy, sad and fear. We also notice that several different AUs may cover the same or similar area. This is attributed to that different AUs may share the same or similar muscle distributions. For instance, all of AU1, AU2 and AU4 are involved with the forehead, while all of AU14, AU15, AU17, AU23 and AU24 are closely related to the mouth area. Besides, considering the existence of potential mutual relations among different AUs, the masks learned merely supervised by AU labels may include other facial regions of correlated AUs. For instance, the masks of AU1, AU2, and AU4 not only cover the regions of the brows but also contain the mouth area.
Mask Features Analysis. To validate the importance of mask learning, we visualize the AU embeddings of the AU classifiers in the GTLE-Net and w/o , respectively. We sample one thousand images and obtain the t-SNE visualization of their AU embeddings.
As illustrated in Figure 6, each color refers to a class of AU presentations. Obviously, the presentations learned with mask(GTLE-Net) are more compact and have better distribution than that generated by model without mask(w/o ). Moreover, all the AU embeddings learned with mask(GTLE-Net) tend to gather into two clusters(occurred and absent samples). By contrast, the presentations generated without the mask extractor(w/o ) display like a long strip and are not divided into two clusters for the cases of AU2, AU4, AU15, AU17 and AU23. As a result, these AUs cannot be well distinguished. This validates that the proposed LAM of our framework is effective to capture discriminative local features that are beneficial to AU recognition.
5 Discussion and Conclusion
Inspired by the correlation between expression representation learning and AU detection, in consideration of the local properties of AUs, in this paper we propose a novel framework for AU detection by learning both expression representation similarity and facial attention. The expression representation learning aims at obtaining powerful and robust latent features that are sensitive to subtle expression transformations. On the other hand, to capture advanced local features concentrated on AU-related areas, we design a local AU feature module(LAM) to learn local attention maps indicating spatial significance and extract local AU representations.
Extensive experimental comparisons with state-of-the-art methods validate that our proposed framework can measurably outperform all of them on the widely-used benchmark datasets (BP4D, DISFA, and BP4D+) for the AU detection task. Our ablation studies not only demonstrate the effectiveness of the prior knowledge from expression similarity learning, but also verifies the validity of our proposed local AU features module(LAM). Through visual analysis, we show that our method can produce highly AU-related attention maps that are sensitive to local facial muscle motions, and that our attention mechanism module can learn compact and discriminative AU embeddings. In addition, we demonstrate the generalization capability of our method through Cross-datasets validation experiments.
While our method achieves superior results on three widely-used datasets, it does not work well for data in the wild. Besides, due to the limitations of AU annotation, it can just detect a part of AUs in FACS. Finally, our method does not explicitly consider the correlation between AUs, which may influence the eventual performance.
Besides AU detection, our method can be potentially extended for other facial analysis applications such as expression recognition and micro-expression detection. In the future, we plan to explore these research directions by extending the current framework.
References
- [1] P. Ekman and W. Friesen, “Facial action coding system: A technique for the measurement of facial movement,” Consulting Psychologists Press Palo Alto, vol. 12, 01 1978.
- [2] M. Pantic and L. Rothkrantz, “Facial action recognition for facial expression analysis from static face images,” Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on, vol. 34, pp. 1449 – 1461, 07 2004.
- [3] G. Zhao and X. Li, “Automatic micro-expression analysis: Open challenges,” Frontiers in Psychology, vol. 10, 08 2019.
- [4] D. R. Rubinow and R. M. Post, “Impaired recognition of affect in facial expression in depressed patients,” Biological psychiatry, vol. 31, no. 9, pp. 947–953, 1992.
- [5] R. Zhi, M. Liu, and D. Zhang, “A comprehensive survey on automatic facial action unit analysis,” The Visual Computer, vol. 36, no. 5, pp. 1067–1093, 2020.
- [6] G. Peng and S. Wang, “Weakly supervised facial action unit recognition through adversarial training,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 2188–2196.
- [7] X. Niu, H. Han, S. Shan, and X. Chen, “Multi-label co-regularization for semi-supervised facial action unit recognition,” Advances in neural information processing systems, vol. 32, 2019.
- [8] X. Zhang, L. Yin, J. F. Cohn, S. Canavan, M. Reale, A. Horowitz, P. Liu, and J. M. Girard, “Bp4d-spontaneous: a high-resolution spontaneous 3d dynamic facial expression database,” Image and Vision Computing, vol. 32, no. 10, pp. 692–706, 2014.
- [9] S. M. Mavadati, M. H. Mahoor, K. Bartlett, P. Trinh, and J. F. Cohn, “Disfa: A spontaneous facial action intensity database,” IEEE Transactions on Affective Computing, vol. 4, no. 2, pp. 151–160, 2013.
- [10] Z. Zhang, J. M. Girard, Y. Wu, X. Zhang, P. Liu, U. Ciftci, S. Canavan, M. Reale, A. Horowitz, H. Yang et al., “Multimodal spontaneous emotion corpus for human behavior analysis,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 3438–3446.
- [11] X. Xiang and T. D. Tran, “Linear disentangled representation learning for facial actions,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 28, no. 12, pp. 3539–3544, 2017.
- [12] X. Niu, H. Han, S. Yang, Y. Huang, and S. Shan, “Local relationship learning with person-specific shape regularization for facial action unit detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 11 917–11 926.
- [13] Y. Chen, D. Chen, T. Wang, Y. Wang, and Y. Liang, “Causal intervention for subject-deconfounded facial action unit recognition,” arXiv preprint arXiv:2204.07935, 2022.
- [14] R. Vemulapalli and A. Agarwala, “A compact embedding for facial expression similarity,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 5683–5692.
- [15] A. Mollahosseini, B. Hasani, and M. H. Mahoor, “Affectnet: A database for facial expression, valence, and arousal computing in the wild,” IEEE Transactions on Affective Computing, vol. 10, no. 1, pp. 18–31, 2017.
- [16] Z. Cui, T. Song, Y. Wang, and Q. Ji, “Knowledge augmented deep neural networks for joint facial expression and action unit recognition,” Advances in Neural Information Processing Systems, vol. 33, 2020.
- [17] M. Liu, S. Li, S. Shan, and X. Chen, “Au-inspired deep networks for facial expression feature learning,” Neurocomputing, vol. 159, pp. 126–136, 2015.
- [18] E. Granger, P. Cardinal et al., “Weakly supervised learning for facial behavior analysis: A review,” arXiv preprint arXiv:2101.09858, 2021.
- [19] L. Zhong, Q. Liu, P. Yang, J. Huang, and D. N. Metaxas, “Learning multiscale active facial patches for expression analysis,” IEEE Transactions on Cybernetics, vol. 45, no. 8, pp. 1499–1510, 2015.
- [20] P. Liu, J. T. Zhou, I. W.-H. Tsang, Z. Meng, S. Han, and Y. Tong, “Feature disentangling machine-a novel approach of feature selection and disentangling in facial expression analysis,” in European Conference on Computer Vision. Springer, 2014, pp. 151–166.
- [21] I. Onal Ertugrul, L. Yang, L. A. Jeni, and J. F. Cohn, “D-pattnet: Dynamic patch-attentive deep network for action unit detection,” Frontiers in computer science, vol. 1, p. 11, 2019.
- [22] K. Zhao, W.-S. Chu, F. De la Torre, J. F. Cohn, and H. Zhang, “Joint patch and multi-label learning for facial action unit detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 2207–2216.
- [23] Z. Liu, J. Dong, C. Zhang, L. Wang, and J. Dang, “Relation modeling with graph convolutional networks for facial action unit detection,” in International Conference on Multimedia Modeling. Springer, 2020, pp. 489–501.
- [24] Z. Shao, Z. Liu, J. Cai, Y. Wu, and L. Ma, “Facial action unit detection using attention and relation learning,” IEEE Transactions on Affective Computing, 2019.
- [25] W. Li, F. Abtahi, Z. Zhu, and L. Yin, “Eac-net: Deep nets with enhancing and cropping for facial action unit detection,” IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 11, pp. 2583–2596, 2018.
- [26] Z. Shao, Z. Liu, J. Cai, and L. Ma, “Jaa-net: joint facial action unit detection and face alignment via adaptive attention,” International Journal of Computer Vision, vol. 129, no. 2, pp. 321–340, 2021.
- [27] G. M. Jacob and B. Stenger, “Facial action unit detection with transformers,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 7680–7689.
- [28] H. Yang, L. Yin, Y. Zhou, and J. Gu, “Exploiting semantic embedding and visual feature for facial action unit detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 10 482–10 491.
- [29] C. Ma, L. Chen, and J. Yong, “Au r-cnn: Encoding expert prior knowledge into r-cnn for action unit detection,” neurocomputing, vol. 355, pp. 35–47, 2019.
- [30] K. Zhao, W. S. Chu, F. De la Torre, J. F. Cohn, and H. Zhang, “Joint patch and multi-label learning for facial action unit and holistic expression recognition,” IEEE Transactions on Image Processing, vol. 25, no. 8, pp. 3931–3946, 2016.
- [31] K. Zhao, W.-S. Chu, and A. M. Martinez, “Learning facial action units from web images with scalable weakly supervised clustering,” in Proceedings of the IEEE Conference on computer vision and pattern recognition, 2018, pp. 2090–2099.
- [32] I. Ö. Ertugrul, L. A. Jeni, and J. F. Cohn, “Pattnet: Patch-attentive deep network for action unit detection.” in BMVC, 2019, p. 114.
- [33] Y. Chang and S. Wang, “Knowledge-driven self-supervised representation learning for facial action unit recognition,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 20 417–20 426.
- [34] K. Wang, X. Peng, J. Yang, D. Meng, and Y. Qiao, “Region attention networks for pose and occlusion robust facial expression recognition,” IEEE Transactions on Image Processing, vol. 29, pp. 4057–4069, 2020.
- [35] Z. Zhao, Q. Liu, and F. Zhou, “Robust lightweight facial expression recognition network with label distribution training,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 4, 2021, pp. 3510–3519.
- [36] P. Paysan, R. Knothe, B. Amberg, S. Romdhani, and T. Vetter, “A 3d face model for pose and illumination invariant face recognition,” in IEEE international conference on advanced video and signal based surveillance, 2009, pp. 296–301.
- [37] J. Booth, A. Roussos, A. Ponniah, D. Dunaway, and S. Zafeiriou, “Large scale 3d morphable models,” International Journal of Computer Vision, vol. 126, no. 2-4, pp. 233–254, 2018.
- [38] L. Li, S. Wang, Z. Zhang, Y. Ding, Y. Zheng, X. Yu, and C. Fan, “Write-a-speaker: Text-based emotional and rhythmic talking-head generation,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 3, 2021, pp. 1911–1920.
- [39] Z. Zhang, L. Li, Y. Ding, and C. Fan, “Flow-guided one-shot talking face generation with a high-resolution audio-visual dataset,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 3661–3670.
- [40] G. Yao, Y. Yuan, T. Shao, S. Li, S. Liu, Y. Liu, M. Wang, and K. Zhou, “One-shot face reenactment using appearance adaptive normalization,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 4, 2021, pp. 3172–3180.
- [41] H. Kim, P. Garrido, A. Tewari, W. Xu, J. Thies, M. Nießner, P. Pérez, C. Richardt, M. Zollhöfer, and C. Theobalt, “Deep video portraits,” TOG, vol. 37, no. 4, pp. 1–14, 2018.
- [42] Z. Geng, C. Cao, and S. Tulyakov, “3d guided fine-grained face manipulation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 9821–9830.
- [43] W. Zhang, X. Ji, K. Chen, Y. Ding, and C. Fan, “Learning a facial expression embedding disentangled from identity,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 6759–6768.
- [44] C. Szegedy, S. Ioffe, V. Vanhoucke, and A. A. Alemi, “Inception-v4, inception-resnet and the impact of residual connections on learning,” in Thirty-first AAAI conference on artificial intelligence, 2017.
- [45] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- [46] ——, “Delving deep into rectifiers: Surpassing human-level performance on imagenet classification,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 1026–1034.
- [47] R.-E. Fan, K.-W. Chang, C.-J. Hsieh, X.-R. Wang, and C.-J. Lin, “Liblinear: A library for large linear classification,” J. Mach. Learn. Res., vol. 9, pp. 1871–1874, 2008.
- [48] K. Zhao, W.-S. Chu, and H. Zhang, “Deep region and multi-label learning for facial action unit detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 3391–3399.
- [49] L. Zhong, Q. Liu, P. Yang, J. Huang, and D. Metaxas, “Learning multiscale active facial patches for expression analysis,” IEEE Transactions on Cybernetics, vol. 45, no. 8, pp. 1499–1510, 8 2015.
- [50] C. Corneanu, M. Madadi, and S. Escalera, “Deep structure inference network for facial action unit recognition,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 298–313.
- [51] G. Li, X. Zhu, Y. Zeng, Q. Wang, and L. Lin, “Semantic relationships guided representation learning for facial action unit recognition,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, 2019, pp. 8594–8601.
- [52] Z.-M. Chen, X.-S. Wei, P. Wang, and Y. Guo, “Multi-label image recognition with graph convolutional networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 5177–5186.
- [53] R. You, Z. Guo, L. Cui, X. Long, Y. Bao, and S. Wen, “Cross-modality attention with semantic graph embedding for multi-label classification,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 07, 2020, pp. 12 709–12 716.
- [54] L. Yang, I. O. Ertugrul, J. F. Cohn, Z. Hammal, D. Jiang, and H. Sahli, “Facs3d-net: 3d convolution based spatiotemporal representation for action unit detection,” in 2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII), 2019, pp. 538–544.
- [55] T. Song, L. Chen, W. Zheng, and Q. Ji, “Uncertain graph neural networks for facial action unit detection,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 1, 2021.
- [56] T. Song, Z. Cui, W. Zheng, and Q. Ji, “Hybrid message passing with performance-driven structures for facial action unit detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 6267–6276.
- [57] Y. Chen, D. Chen, Y. Wang, T. Wang, and Y. Liang, “Cafgraph: Context-aware facial multi-graph representation for facial action unit recognition,” in Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 1029–1037.
- [58] L. A. Jeni, J. F. Cohn, and F. De La Torre, “Facing imbalanced data–recommendations for the use of performance metrics,” in 2013 Humaine association conference on affective computing and intelligent interaction, 2013, pp. 245–251.
- [59] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” Advances in neural information processing systems, vol. 25, 2012.
![]() |
Rudong An received the B.S. degree and M.S. degree in college of Biomedical Engineering and Instrument Science from Zhejiang University, Hangzhou, China, in 2018 and 2021. He is currently an artificial intelligence researcher at Netease Fuxi AI Lab. His interests include deep learning, action unit detection, facial expression analysis and facial animation generation. |
![]() |
Wei Zhang received the B.E. degree in communication engineering from Nanjing University of Posts and Telecommunications, Jiangsu, China in 2017 and M.S. degree in electronic and information engineering from Zhejiang University, Zhejiang, China in 2020. She is currently a research scientist working with Netease Fuxi AI Lab, Hangzhou, China. Her current research interests include computer vision, expression embedding and facial affective analysis. |
![]() |
Hao Zeng received the B.E. degree in computer software engineering and the M.S. degree in computer science and technology from Huazhong University of Science and Technology, Hubei, China, in 2017 and 2020, respectively. He is currently a researcher in Netease Fuxi AI Lab, Hangzhou, China. His current research interests include deep learning, computer vision and face generation. |
![]() |
Wei Chen received the B.S. degree in information engineering from the National University of Defense Technology, Changsha, China, and received M.S. degree in software engineering from Hebei University, Baoding, China. He is currently an engineer of Hebei Agricultural University, Baoding, China. His interests include software engineering, intelligent optimization algorithm, and deep learning. |
![]() |
Zhigang Deng is Moores Professor of Computer Science at University of Houston, Texas, USA. His research interests include computer graphics, computer animation, virtual humans, human computer conversation, and robotics. He earned his Ph.D. in Computer Science at the Department of Computer Science at the University of Southern California in 2006. Prior that, he also completed B.S. degree in Mathematics from Xiamen University (China), and M.S. in Computer Science from Peking University (China). Besides serving as the conference general/program co-chair for CASA 2014, SCA 2015, and MIG 2022, he has been an Associate Editor for IEEE Transactions on Visualization and Computer Graphics, Computer Graphics Forum, Computer Animation and Virtual Worlds Journal, etc. He is a distinguished member of ACM and a senior member of IEEE. |
![]() |
Yu Ding is currently an artificial intelligence expert, leading the virtual human group at Netease Fuxi AI Lab, Hangzhou, China. His research interests include deep learning, image and video processing, talking face generation, animation generation, facial expression recognition, multimodal computing, affective computing, nonverbal communication (face, gaze, and gesture), and embodied conversational agent. He received Ph.D. degree in Computer Science (2014) at Telecom Paristech in Paris (France), M.S. degree in Computer Science at Pierre and Marie Curie University (France), and B.S. degree in Automation at Xiamen University (China). |