This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

ICE-GAN: Identity-aware and Capsule-Enhanced GAN with Graph-based Reasoning for Micro-Expression Recognition and Synthesis

Jianhui Yu1, Chaoyi Zhang1, Yang Song2, Weidong Cai1 1School of Computer Science, University of Sydney, Australia
{jianhui.yu, chaoyi.zhang, tom.cai}@sydney.edu.au
2School of Computer Science and Engineering, University of New South Wales, Australia
[email protected]
Abstract

Micro-expressions are reflections of people’s true feelings and motives, which attract an increasing number of researchers into the study of automatic facial micro-expression recognition. The short detection window, the subtle facial muscle movements, and the limited training samples make micro-expression recognition challenging. To this end, we propose a novel Identity-aware and Capsule-Enhanced Generative Adversarial Network with graph-based reasoning (ICE-GAN), introducing micro-expression synthesis as an auxiliary task to assist recognition. The generator produces synthetic faces with controllable micro-expressions and identity-aware features, whose long-ranged dependencies are captured through the graph reasoning module (GRM), and the discriminator detects the image authenticity and expression classes. Our ICE-GAN was evaluated on Micro-Expression Grand Challenge 2019 (MEGC2019) with a significant improvement (12.9%) over the winner and surpassed other state-of-the-art methods.

Index Terms:
micro-expression synthesis; micro-expression recognition; generative adversarial network; capsule network; graph relation learning

I Introduction

Micro-expressions (MEs) display unconscious feelings that can be hardly perceived by untrained observers, making it a challenging pattern recognition task. However, it has attracted an increasing number of researchers into the study of micro-expression recognition (MER) due to its practical applications in areas such as [1] and [2]. There have already been many successful works proposed for general expression recognition, but the domain of MER is poorly-explored mainly because no large-scale ME datasets are available to support extensive MER studies. Therefore, we introduce micro-expression synthesis (MES) as an auxiliary task to intentionally increase the number of ME samples.

Before the application of deep learning algorithms, hand-engineered methods for recognizing MEs are used according to [1]. Facial Action Coding System (FACS) [3] is applied to recognize facial expressions, which pays attention to muscles that produce the expressions and measures the movement with the help of action units (AUs). Two systems are further developed: the Micro Expression Training Tool (METT) and Subtle Expression Training Tool (SETT) [4]. However, the best classification accuracy achieved by METT/SETT is still low because the result is heavily affected by humans, making the detection unconvincing and unstable. Local binary pattern (LBP) and local quantized pattern (LQP) are later developed [1], and LBP with three orthogonal planes (LBP-TOP [5]) has shown superiority in processing facial images. However, these geometry-based methods rely heavily on the proposed images and can be easily affected by global changes.

With the development of deep learning technologies, especially convolutional neural networks (CNNs), lots of works have been proposed based on data-driven approaches for MER [6, 7, 8]. ELRCN [6] adopts the model architecture of [9] with enriched features to capture subtle facial movements from frame sequences. Moreover, methods utilize the optical flow to enhance the model performance. STSTNet [10] extracts three optical features for lightweight network construction. Dual-Inception [11] learns the facial MEs with the help of horizontal and vertical optical flow components. ME-Recognizer [7] obtains optical features encode the subtle face motion with domain adaptation, which achieves the 1st1^{st} place in MEGC2019 [12]. However, CNN-based methods are invariant to translations and do not encode positional relations, and orientation information of different facial entities is also ignored. Furthermore, due to the quantity limit of ME samples, the model performance is heavily constrained.

Capsule network [13] is translation equivariant, which is adopted for MER in [8] to distinguish meaningful facial patterns. In addition, the size limitation of current ME datasets (i.e., SMIC [14], CASME II [15], and SAMM [16]) puts a constraint on deep learning methods that can hardly derive benefits from small-scale datasets. However, this can be resolved by adopting generative adversarial network (GAN) [17] and the variants [18, 19, 20], which have shown significant generative capabilities in various vision application fields [21, 22], so that GANs can naturally help to expand the size of micro-expression datasets.

Relation learning can be crucial for the general visual analysis [23], and capturing global relations during the training process is quite beneficial [24]. We propose to learn relations globally on feature channels to capture the long-range channel interdependencies as in [25], with the help of self-attention mechanism [26]. Moreover, graph convolution networks (GCNs) [27] are proposed and adopted initially by [28] for feature representation and textual information learning, where additional information such as AUs is needed to construct the graph. However, we propose to construct graphs based on channel features, so the amount of needed information is heavily reduced and global relations between different facial attributes can be learned, where node features and edge features are learned during training.

The main contributions of our work can be summarized as 3-fold: (1) We design an identity-aware generator that produces controllable expressions based on side information and generates realistic faces with a graph reasoning module, which models channel interdependencies via attention mechanism to fully explore global relations between facial encodings on local responses. (2) We introduce a multi-tasking capsule-enhanced discriminator to distinguish image authenticity and predict micro-expression labels, with position-insensitive issues alleviated by the capsule-based algorithm to improve the MER accuracy. (3) We thereby develop an ICE-GAN framework for identity-aware MES with graph-based reasoning and capsule-enhanced MER, which outperforms the winner of the MEGC2019 Challenge benchmark and recently proposed methods.

II Related Work

II-A Generative Adversarial Network

Existing databases for MER constrain the development of data-driven deep learning methods because overfitting is considered a major problem. To enlarge the current dataset, GANs can generate unseen images by modeling the training data distribution, where distribution mappings are learned by the generator and the authenticity of inputs is determined by the discriminator. So far there have been tremendous extensions of GANs [20, 21, 18], and applications of GANs to MEs are also explored [29, 30]. For instance, optical flow images are generated via GAN to enlarge the database in [29], and real facial images are produced based on AU intensity in [30]. However, our method is inspired by Auxiliary Classifier GAN (ACGAN) [19] to generate unseen MEs based on image features with controllable categories, which can be of high quality and high discriminability.

II-B Graph-based Reasoning

Global relations between different nodes are initially modeled by conditional random fields [31]. Recently, graph convolutional networks (GCNs) [27] are proposed and adopted in [23, 32, 33], and [34], where relational information is reasoned over graphs, and contextual interactions over textual or visual contents are captured. Besides, graph-based reasoning methods are applied in the study field of MEs. [30, 35, 36] facilitate the learning of subtle facial movements from graphs that are established on AU nodes, and structural relations are captured for geometric reasoning. However, instead of directly constructing graphs over raw data inputs, we propose to explore global relations over intermediate feature channels for representative relation learning. Inspired by SENet [25] which applies attention mechanism [26] to model the interactions between input features and channel features where spatial information is pooled, we employ self-attention within feature channels to save low-level spatial features, resulting in an attentive mapping that contains dependencies between facial attributes. Moreover, different from [30] which implements MER with supplementary information such as AUs, our work can directly learn the relational information of different face regions, reducing the amount of information required.

II-C Capsule Network

Different from conventional CNNs that are translation invariant, the capsule network is translation equivariant, which presents a competitive learning capacity considering the relative pose and position information of object entities in the image. The reasonable layouts of different face parts do matter during MER. Ertugrul et al. [37] manage to encode face poses and AUs at different view angles with the help of a single capsule. LaLonde et al. [38] develop a deconvolutional capsule network with U-net architecture [39] which achieves a good result in object segmentation. [8] is the first work that manages to use capsule-based architecture for MER, whose model performance is better than the LBP-TOP baseline and several CNN models. In contrast, we propose our discriminator design which is enhanced by capsule network to implement multiple tasks: to check whether the input image is real or fake and to predict the ME labels.

Refer to caption
Figure 1: Architecture of the proposed ICE-GAN framework for MES and MER tasks, where GencG_{enc} and GdecG_{dec} represent the encoder and decoder part of generator, and DD denotes the capsule discriminator.

III Methods

The overall architecture is shown in Fig. 1. The onset image XonX_{on} is regarded as the neutral face with the lowest expressive intensity. U-net like generator is used to modulate XonX_{on} with side information (i.e., random noise zz and class label cc) and produce the output XsynX_{syn} with a desired class, preserving the identity knowledge. The real apex expressive faces XapexX_{apex} and XsynX_{syn} are adopted to train our capsule-enhanced discriminator, which is a multi-tasking component for authenticity checking between XapexX_{apex} and XsynX_{syn}, and for expression classification. More details are elaborated in the following sections.

III-A Micro-Expression Synthesis via Identity-aware Generator

We propose an identity-aware generator GG based on the encoder-decoder structure shown in Fig. 2, which can preserve the identity feature and produce outputs with preferable classes during MES, with the help of side information. The encoding procedure explores facial attributes from onset neutral inputs XonX_{on} via a series of hierarchical convolutional layers, resulting in intermediate feature maps fi:1,2,,6encCi×Hi×Wif_{i:1,2,...,6}^{enc}\in\mathcal{R}^{C_{i}\times H_{i}\times W_{i}} at the ii-th convolutional layer, where CiC_{i}, HiH_{i}, and WiW_{i} denote the channel, height, and width of the feature map, respectively.

Furthermore, we decide to utilize fiencf_{i}^{enc} by processing and propagating the intermediate information from the encoder GencG_{enc} to decoder GdecG_{dec} for more realistic image generation, as these local features contain trivial knowledge such as 2D face geometries and non-trivial knowledge that is useful for the expression generation. So, we propose to use a reasoning module based on graphs, namely graph reasoning module (GRM), to capture facial part relations to reduce artifacts.

Instead of leveraging skip connections to directly transfer the multi-scale spatial information, we design to firstly flatten the spatial dimension and directly apply self-attention on feature channels to learn a channel map. Moreover, we propose to use graph convolution to better reason the relationships between feature channels by treating the channel map as a graph, which is named as the channel graph, so that global interdependencies between intermediate facial attributes can be captured from local feature responses.

The detailed architecture is displayed on the yellow box of Fig. 2. Based on the feature map fiencCi×Hi×Wif^{enc}_{i}\in\mathcal{R}^{C_{i}\times H_{i}\times W_{i}} after convolution, we can construct a spatial graph with the node number Ns=Hi×WiN_{s}=H_{i}\times W_{i} and node feature Cs=CiC_{s}=C_{i}. Inspired by [32], supernodes with richer expressiveness are formed by a transformation function T()T(\cdot) to generate a new feature map f^iencCi×Ns^\hat{f}_{i}^{enc}\in\mathcal{R}^{C_{i}\times\hat{N_{s}}}, and empirically, T()T(\cdot) is chosen as a convolutional operation. To learn global relations between different facial encodings, self-attention is applied to calculate a similarity mapping of features based on f^ienc\hat{f}_{i}^{enc}, which can be formulated as:

=ϕ(f^ienc)θ(f^ienc)T,\mathcal{M}=\phi(\hat{f}_{i}^{enc})\theta(\hat{f}_{i}^{enc})^{T}, (1)

where both ϕ()\phi(\cdot) and θ()\theta(\cdot) are linear mappings. We take the similarity mapping Ci×Ci\mathcal{M}\in\mathcal{R}^{C_{i}\times C_{i}} as a channel graph, which has a set of node features 𝐦={mj:1,2,..,Ci1×Ci}\mathbf{m}=\{m_{j:1,2,..,C_{i}}\in\mathcal{R}^{1\times C_{i}}\}. In our case, we define \mathcal{M} as an undirected and fully connected graph.

To reason over the whole channel graph, we utilize a GCN [27] to learn edge features and update node features. As the initial edge feature between two nodes is unknown and undefined, we propose to set up a self-learning adjacency matrix ACi×CiA\in\mathcal{R}^{C_{i}\times C_{i}}, which can be randomly initialized following the normal distribution and updated during back-propagation. The whole process can be expressed as:

^=σ((A+I)W),\hat{\mathcal{M}}=\sigma((A+I)\mathcal{M}W), (2)

where σ\sigma is an activation function, identity matrix II is added to construct self-looping, and WCi×CiW\in\mathcal{R}^{C_{i}\times C_{i}} and ACi×CiA\in\mathcal{R}^{C_{i}\times C_{i}} denote the weight and adjacency matrices, which are both learnable.

We then project the updated graph ^\hat{\mathcal{M}} back to the same feature space of fiencf_{i}^{enc} through an inverse projection function T1()T^{-1}(\cdot). Lastly, the residual connection is employed to compensate the high-level semantics relations with the low-level information. The final output gidecg^{dec}_{i} of the graph reasoning module is represented as gidec=fiencT1(^)g^{dec}_{i}=f^{enc}_{i}\oplus T^{-1}(\hat{\mathcal{M}}).

The last layer of GencG_{enc} is the bottleneck layer that learns a compressed representation of the input data, termed as identity-aware embedding e320×1×1e\in\mathcal{R}^{320\times 1\times 1}. As a controlling term for expression generation, the class label cc is concatenated with ee and random noise zz to form a synthesis seed s423×1×1s\in\mathcal{R}^{423\times 1\times 1}, which is fed into the decoding part where the spatial dimensions of feature maps are expanded and feature channels are reduced. Thus the decoding feature maps fi:1,2,,6decf^{dec}_{i:1,2,...,6} can be obtained with the help of deconvolution operation h()h(\cdot) [40]. Moreover, gidecg^{dec}_{i} will be further concatenated with fidecf^{dec}_{i} so the learned relational information from latent feature space can be leveraged to improve synthetic image quality. The ii-th upper layer feature map (for i[2,6]i\in[2,6]) can be represented as:

fi1dec=h(gidecfidec).f^{dec}_{i-1}=h(g^{dec}_{i}\oplus f^{dec}_{i}). (3)
Refer to caption
Figure 2: The encoder-decoder framework of our proposed generator GG, where the detailed implementations of the graph construction and global reasoning are illustrated within the yellow box.

III-B Micro-Expression Recognition via Capsule-Enhanced Discriminator

Unlike CNNs operating over single scalars, capsule network attends to vectors, of which lengths are used to represent the existence probabilities of each entity in a given image. We design a multi-tasking discriminator for sample authenticity checking and ME label classification enhanced by capsule network, which enables the learning of richer visual expressions and more sensitive to the geometric encoding of relative positions and poses of entities than conventional CNNs, dubbed a capsule-enhanced discriminator.

Fig. 3 presents the detailed architecture. The facial attributes are encoded via PatchGAN for faster learning, which are then fed into PrimaryCaps to encapsulate the information at a lower level. Vectors generated from PrimaryCaps are coupled and used to activate the capsules in the next layer. Two following capsules, namely AdvCaps and ExpCaps, are designed for two separate sub-tasks: (1) AdvCaps distinguishes the real expressive images XapexX_{apex} from the synthetic ones XsynX_{syn}, and (2) ExpCaps predicts the corresponding micro-expression labels of input images. Furthermore, a reconstruction network [13] is desired for a performance gain by regularizing the training of ExpCaps.

Refer to caption
Figure 3: The capsule-enhanced discriminator DD of our proposed method, where NprimN_{prim} denotes the number of primary capsules, dprimd_{prim}, dexpd_{exp}, and dadvd_{adv} represent the dimension of each PrimaryCaps, ExpCaps, and AdvCaps, respectively.

III-C Model Objective

Generator Objective. A two-term identity-preserving loss LipL_{ip} is adopted to capture the identity information embedded in XonX_{on} (via GencG_{enc}) and thus to synthesize distinctive samples XsynX_{syn} (via GdecG_{dec}). One term of LipL_{ip} is selected as the pixel-wise reconstruction loss LpixelL_{pixel} to improve the generated image quality. We empirically use L1 penalty over L2 penalty to establish direct supervision on XsynX_{syn}. Meanwhile, we preserve a perceptual similarity between XonX_{on} and XsynX_{syn} during MES as in [41] by introducing a perceptual loss LperL_{per}, which is adopted as the second term in LipL_{ip} to preserve the facial styles with regards to different subjects. With the help of a cost network which is usually implemented as a pre-trained CNN, LperL_{per} can be easily estimated and minimized over high-level feature representations associated with XonX_{on} and XsynX_{syn}. The overall representation of Lip(G)L_{ip}(G) can be described as:

Lip(G)=Lpixel(G)+αLper(G).L_{ip}(G)=L_{pixel}(G)+\alpha L_{per}(G). (4)

Discriminator Objective. The training of the capsule-enhanced discriminator is optimized by the margin loss LmarginL_{margin} as suggested in [13], which can enlarge the feature distance of different facial expressions. LmarginL_{margin} can be obtained via:

Lmargin(D)=Tkmax(0,m+𝒗𝒌)2+λk(1Tk)max(0,𝒗𝒌m)2,\begin{split}L_{margin}(D)\;=\;&T_{k}max(0,m^{+}-\lVert\boldsymbol{v_{k}}\rVert)^{2}\;+\\ \;&\lambda_{k}(1-T_{k})max(0,\lVert\boldsymbol{v_{k}}\rVert-m^{-})^{2},\end{split} (5)

where TkT_{k} = 1 if expression class kk exists otherwise 0, while m+m^{+} and mm^{-} are the upper and lower margins and vkv_{k} is the vector output of capsules that being activated to class kk. Moreover, the mean square error is employed as the loss function LrecL_{rec} in the reconstruction network to regularize the training procedure. Hence, the classification-related loss LclsL_{cls} for our capsule-enhanced discriminator is summed as:

Lcls(D)=Lmargin(D)+βLrec(D).L_{cls}(D)=L_{margin}(D)+\beta L_{rec}(D). (6)

Overall Objective. As we treat GAN as our baseline, the optimization process between the generator and discriminator is described as a min-max game [18]. The learning objective LganL_{gan} can be formulated as:

Lgan=minGmaxDV(D,G)=Expx[log(D(x|y))]+Ezpz[1D(G(z|c))],\begin{split}L_{gan}=\mathop{min}\limits_{G}\mathop{max}\limits_{D}V(D,G)\;=\;&E_{x\sim p_{x}}[log(D(x|y))]\;+\\ \;&E_{z\sim p_{z}}[1-D(G(z|c))],\end{split} (7)

where xx and yy indicate the real images and the real labels, while zz and cc denote the random noise and fake labels.

Overall, the total loss function of our proposed ICE-GAN can be summarized as:

Ltot=λadvLgan(D,G)+λmesLip(G)+λmerLcls(D),L_{tot}=\lambda_{adv}L_{gan}(D,G)+\lambda_{mes}L_{ip}(G)+\lambda_{mer}L_{cls}(D), (8)

where λadv\lambda_{adv}, λmes\lambda_{mes}, and λmer\lambda_{mer} are multi-tasking weight parameters of the proposed tasks including GAN optimization, identity-aware MES, and capsule-enhanced MER.

IV Experiments

In this section, we compared ICE-GAN with other state-of-the-art methods quantitatively for MER based on two evaluation protocols: cross-database evaluation and single-database evaluation. Additionally, qualitative analyses of synthetic images were reported on MES. Moreover, extensive experiments on our module designs were implemented for ablation studies.

IV-A Dataset Descriptions

Evaluations were conducted on the MEGC2019 cross-database benchmark, which consists of three publicly available ME datasets (SMIC [14], CASME II [15], and SAMM [16]), with a total of 442 ME video sequences (164, 145, and 133 clips collected from SMIC, CASME II, and SAMM, respectively) captured from 68 subjects (16, 24, and 28 subjects from SMIC, CASME II, and SAMM, respectively). Frames in all 3 datasets are categorized into 3 classes of positive, negative, and surprise with the Leave-One-Subject-Out (LOSO) validation method. LOSO is conducted for subject-independent evaluation, i.e., the evaluation is repeated each time for all subjects accordingly, until each subject is split alone as the 1-subject testing dataset and the remaining subjects as the training dataset. In our case, we firstly conducted experiments on all 3 datasets for cross-database evaluation (CDE) based on the Unweighted F1-score (UF1) and Unweighted Average Recall (UAR), making a fair performance comparison as in [12]. We then evaluated model performance only on CASME II and SAMM for single-database evaluation (SDE) using F1-score following [30].

IV-B Implementation Details

In this section, we reveal detailed implementations of data preprocessing and model parameter settings for our design.

Data Preprocessing. We firstly cropped out face regions by conducting facial landmark detection repeatedly to locate and refine the positions of 68 facial landmarks [42], ending up with a bounding box decided by the refined 68 landmarks with toolkit 111https://pypi.org/project/face-recognition/. We then followed [8] to find the onset and apex frames in SMIC because annotated positions of onset and apex frames are only available in CASME II and SAMM. Moreover, we chose the neighboring four images around the apex image for each subject to augment the existing data, where we assumed that they all have the same expressive intensity. Finally, these cropped face images were resized to 128×\times128 in grayscale.

TABLE I: MER results on the MEGC2019 benchmark and separate datasets in terms of UF1 and UAR with LOSO cross-database evaluation.
Method MEGR 2019 SAMM SMIC CASME II
UF1 UAR UF1 UAR UF1 UAR UF1 UAR
LBP-TOP [5] 0.588 0.578 0.395 0.410 0.200 0.528 0.702 0.742
CapsuleNet [8] 0.652 0.650 0.620 0.598 0.582 0.587 0.706 0.701
Dual-Inception [11] 0.732 0.727 0.586 0.566 0.664 0.672 0.862 0.856
STSTNet [10] 0.735 0.760 0.658 0.681 0.680 0.701 0.838 0.868
ME-Recognizer [7] 0.788 0.782 0.775 0.715 0.746 0.753 0.829 0.820
ICE-GAN 0.874 0.883 0.879 0.883 0.782 0.801 0.895 0.904

Model Parameters. The multi-tasking weight parameters in (8) are set as follows: λadv\lambda_{adv} = 0.1, λmes\lambda_{mes} = λmer\lambda_{mer} = 1. Two reweighting ratios α\alpha and β\beta are set to 0.1 and 5e-4. We set the noise dimension NzN_{z} to 100 and class label dimension NcN_{c} to 3 as we only have 3 classes. The discriminator includes a 70×\times70 PatchGAN followed by NprimN_{prim} = 8 PrimaryCaps with dimension dprimd_{prim} = 16 and two sub-capsules for classification purposes (dadvd_{adv} = 256 and dexpd_{exp} = 32). m+m^{+}, mm^{-}, and λk\lambda_{k} in (5) are set to 0.9, 0.1, and 0.5. Adam is selected as the optimizer to train our network, with momentums b1b_{1} and b2b_{2} set to 0.9 and 0.999, respectively. The learning rate is initialized as 1e-3 and decays using a cosine annealing schedule. The batch size is set to 16 with 100 training epochs. The end-to-end training procedure of ICE-GAN was implemented in Pytorch with one Nvidia RTX2080Ti GPU.

IV-C Evaluations

Quantitative Results. We firstly evaluated our model performance with state-of-the-art methods from MEGC2019 benchmark [12] using CDE. The results of UF1 and UAR are reported in Table I. It can be observed that ICE-GAN outperforms approaches from MEGC2019 benchmark, where UF1 and UAR scores are improved by 10.9% and 12.9% compared to those of the winner [7].

The baseline (i.e., LBP-TOP) in Table I adopts hand-crafted features and receives a lower result compared to deep learning methods. CapsuleNet [8] utilizes capsule network to achieve an acceptable result, and the remaining works (i.e., [10, 11, 7]) propose their designs based on optical flows. The superior performance of our design can be attributed to the expanded size of the database and the global relations of facial attributes which are learned on channel graphs. Although ME-Recognizer [7] captures spatiotemporal information from facial movements between onset and apex images, the intra-class information is ignored for each subject.

Our model also achieves the highest scores for individual dataset evaluation. For SAMM, ICE-GAN overpasses ME-Recognizer by 13.4% in UF1 and 23.5% in UAR. The performance in CASME II has improved by 8.0% and 10.2% for UF1 and UAR respectively. However, the improvement on SMIC is not as large as the previous two datasets because there is no clear annotation about onset or apex frames, which cannot give representative information about the subject.

We then compared our method with the latest state-of-the-art works following SDE. The model performance of F1-score for 3-category classification is reported in Table II. Compared to MicroAttention [24], since we capture the long-ranged interactions between different facial regions based on graph reasoning, our method achieves a higher F1-score with 8.5% improvement on CASME II and 54.0% on SAMM. MER-GCN [35] and AU-GACN [30] are graph-based methods that reason over AU nodes, while our ICE-GAN achieves a better performance with F1-scores improved by 64.8% and 43.9% compared to AU-GACN, which indicates that our way of constructing graphs over intermediate feature channels can give more representative relational information between facial attributes than directly reason over AU graphs.

Qualitative Results. Furthermore, qualitative results for MES are shown in Fig. 4, where comparisons between synthetic and real images of four different subjects are demonstrated. The results indicate that the generated faces achieve no significant artifacts and are basically at the same level as real samples regarding the authenticity, where light conditions are preserved in the images as well.

Refer to caption
Refer to caption
Figure 4: (a) Comparisons between real samples and generated samples by ICE-GAN for four different subjects. Neutral images are listed for reference. (b) Synthetic images XsynX_{syn} generated by different model designs.
Refer to caption
Figure 5: Norm2-based difference between XsynX_{syn} and XonX_{on}. Green boxes indicate the subtle muscle movements associated with action units.
TABLE II: Comparisons with the latest graph-based and attention-based approaches in terms of F1-score with LOSO single-database evaluation.
Method CASME II SAMM
MicroAttention [24] 0.539 0.402
MER-GCN[35] 0.303 0.283
AU-GACN[30] 0.355 0.433
ICE-GAN 0.585 0.623

To better examine synthetic samples, we implemented Norm2-based difference between XsynX_{syn} and XonX_{on} and focused on AU-related face regions. Fig. 5 presents activated regions shown in red boxes for positive and surprising faces, while negative expressions have more than one kind of MEs and therefore their patterns are difficult to be visualized explicitly. Practically, facial expressions relate to many muscle movements, so multiple AUs can be activated simultaneously. For positive classes, green boxes drawn around the eye area indicate the activation of AU6, and the ones around the lip corner indicate AU12 and AU25. For surprising samples, AU1, AU2, and AU5 are activated, which are all prototypical AUs that are critical for the recognition of positive (i.e., happy) and surprise micro-expressions, according to FACS [3].

IV-D Ablation Studies

Extensive experiments were implemented to validate the component design of ICE-GAN. The following studies were examined on the full dataset including SMIC, CASME II, and SAMM, which was further randomly split into a training set with 48 subjects and a testing set with 20 subjects.

TABLE III: Experimental analyses of GG and GRM designs on the full dataset split in a subject-wise manner. SCSC, SESE, and GRGR represent the skip connection, squeeze-and-excitation, and graph reasoning.
Model DD GdecG_{dec} GencG_{enc} SCSC SESE GRGR UAR UF1
A \checkmark 0.425 0.423
B \checkmark \checkmark 0.651 0.660
C \checkmark \checkmark \checkmark 0.705 0.707
D \checkmark \checkmark \checkmark \checkmark 0.717 0.724
E \checkmark \checkmark \checkmark \checkmark 0.719 0.732
F \checkmark \checkmark \checkmark \checkmark 0.761 0.769

Analysis of Generator. Three models were built to verify the effectiveness of our generator in Table III. Model A is the baseline that just uses a discriminative model for MER. Model B includes a DCGAN-like generator [20], where UAR and UF1 are increased by 53.2% and 56.0% compared to model A, which validates the usefulness of GAN to expand the data size. Model C achieves UAR of 0.705 and UF1 of 0.707 based on an encoder-decoder like structure, which proves the advantage of encoding expressive representations of the input data.

Analysis of Graph Reasoning Module. To validate the excellence of the graph-based reasoning module, we compared its performance with skip connections (model D) and squeeze-and-excitation (SE) module [25] (model E) in Table III. Skip connections simply propagate low-level multi-scale spatial information from encoder to decoder side, however, there is no relation learning or reasoning on latent feature space. Thus, model D can only achieve UAR of 0.717 and UF1 of 0.724. Results shown in model E indicate that SE module helps the model modulate the channel interdependencies, and model capability is increased compared to model D with only skip connections. Moreover, by constructing graphs based on feature channels and learning global relations on local facial features, the model performance has been further improved in the final design (model F). Our model achieves the best performance with UAR of 0.761 and UF1 of 0.769, which indicates that relational information between different facial features reasoned from local responses contributes significantly when generating unseen faces.

Analysis of Synthesis Quality. Moreover, synthetic image qualities from models B, C, and F are visualized in Fig. 4. The differences between models B and C demonstrate that finer identity-related attributes can be preserved via the encoder-decoder architecture. With the help of graph-based global reasoning, more high-frequency signals can be passed smoothly through multi-scale connections (e.g. light condition) and artifacts are largely reduced.

Analysis of Discriminator Design. We then exploited the impact of the dimension dexpd_{exp} of ExpCaps on MER, within a range of [8, 16, 32, 64, 128]. As observed in Table IV, we obtained the best performance when setting dexpd_{exp} to 32, with UAR of 0.761 and UF1 of 0.769. Besides, we compared the capsule-enhanced design with its CNN-based counterparts by replacing the two-layer capsule network with a two-layer CNN, while maintaining the architecture of PatchGAN. The performance of CNN-based discriminator is reported in the first row in Table IV. To make a fair comparison of the difference in terms of model size, we increased the number of neurons in the two-layer CNN and examined the performance of the enlarged CNN-based discriminator (second row) on the MER task. As reported in Table IV, our capsule-enhanced discriminator outperforms its CNN counterparts by a large margin in terms of UAR and UF1, validating that translation equivariance introduced by capsule helps MER.

TABLE IV: Ablation studies on the discriminator design. DCNND_{CNN} and DCED_{CE} denote CNN-based and capsule-enhanced discriminators.
Method UAR UF1 #Params
DCNND_{CNN} 0.356 0.321 6.7 MB
DCNND_{CNN} with comparable size 0.432 0.455 85.6 MB
DCED_{CE} with dexp=8d_{exp}=8 0.730 0.740 85.7 MB
DCED_{CE} with dexp=16d_{exp}=16 0.704 0.723 85.9 MB
DCED_{CE} with dexp=32d_{exp}=32 0.761 0.769 86.4 MB
DCED_{CE} with dexp=64d_{exp}=64 0.726 0.732 87.5 MB
DCED_{CE} with dexp=128d_{exp}=128 0.713 0.730 89.6 MB

V Conclusion

In this work, we propose ICE-GAN which consists of an identity-aware generator with global graph reasoning on local channel graphs for micro-expression synthesis and a capsule-enhanced discriminator for micro-expression recognition. We design a generator that encodes distinguishable facial attributes with side information to control expressions and synthesizes realistic samples by a graph reasoning module, where a channel graph is established via self-attention and global relations between long-ranged facial features are captured. Furthermore, we present a discriminator which improves recognition ability by capturing part-based position-specific face characteristics. Experiments on MEGC2019 datasets indicate the proposed method outperforms the winner by a significant margin, and ablation studies demonstrate our contributions to different module designs.

References

  • [1] W. Merghani, A. K. Davison, and M. H. Yap, “A review on facial micro-expressions analysis: datasets, features and metrics,” arXiv preprint arXiv:1805.02397, 2018.
  • [2] M. Takalkar, M. Xu, Q. Wu, and Z. Chaczko, “A survey: facial micro-expression recognition,” Multimedia Tools and Applications, vol. 77, no. 15, pp. 19301–19325, 2018.
  • [3] J. F. Cohn, Z. Ambadar, and P. Ekman, “Observer-based measurement of facial expression with the facial action coding system,” The handbook of emotion elicitation and assessment, vol. 1, no. 3, pp. 203–221, 2007.
  • [4] P. Ekman, Micro expressions training tool. Emotionsrevealed. com, 2003.
  • [5] G. Zhao and M. Pietikainen, “Dynamic texture recognition using local binary patterns with an application to facial expressions,” IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), vol. 29, no. 6, pp. 915–928, 2007.
  • [6] H.-Q. Khor, J. See, R. C. W. Phan, and W. Lin, “Enriched long-term recurrent convolutional network for facial micro-expression recognition,” in 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG), pp. 667–674, IEEE, 2018.
  • [7] Y. Liu, H. Du, L. Zheng, and T. Gedeon, “A neural micro-expression recognizer,” in 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG), pp. 1–4, IEEE, 2019.
  • [8] N. Van Quang, J. Chun, and T. Tokuyama, “Capsulenet for micro-expression recognition,” in 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG), pp. 1–7, IEEE, 2019.
  • [9] J. Donahue, L. Anne Hendricks, S. Guadarrama, M. Rohrbach, S. Venugopalan, K. Saenko, and T. Darrell, “Long-term recurrent convolutional networks for visual recognition and description,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2625–2634, 2015.
  • [10] S.-T. Liong, Y. Gan, J. See, H.-Q. Khor, and Y.-C. Huang, “Shallow triple stream three-dimensional cnn (ststnet) for micro-expression recognition,” in 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG), pp. 1–5, IEEE, 2019.
  • [11] L. Zhou, Q. Mao, and L. Xue, “Dual-inception network for cross-database micro-expression recognition,” in 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG), pp. 1–5, IEEE, 2019.
  • [12] J. See, M. H. Yap, J. Li, X. Hong, and S.-J. Wang, “Megc 2019–the second facial micro-expressions grand challenge,” in 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG), pp. 1–5, IEEE, 2019.
  • [13] S. Sabour, N. Frosst, and G. E. Hinton, “Dynamic routing between capsules,” in Advances in neural information processing systems (NeurIPS), pp. 3856–3866, 2017.
  • [14] X. Li, T. Pfister, X. Huang, G. Zhao, and M. Pietikäinen, “A spontaneous micro-expression database: Inducement, collection and baseline,” in 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), pp. 1–6, IEEE, 2013.
  • [15] W.-J. Yan, X. Li, S.-J. Wang, G. Zhao, Y.-J. Liu, Y.-H. Chen, and X. Fu, “Casme ii: An improved spontaneous micro-expression database and the baseline evaluation,” PloS one, vol. 9, no. 1, 2014.
  • [16] A. K. Davison, C. Lansley, N. Costen, K. Tan, and M. H. Yap, “Samm: A spontaneous micro-facial movement dataset,” IEEE Transactions on Affective Computing (TAC), vol. 9, no. 1, pp. 116–129, 2016.
  • [17] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in Advances in neural information processing systems (NeurIPS), pp. 2672–2680, 2014.
  • [18] M. Mirza and S. Osindero, “Conditional generative adversarial nets,” arXiv preprint arXiv:1411.1784, 2014.
  • [19] A. Odena, C. Olah, and J. Shlens, “Conditional image synthesis with auxiliary classifier gans,” in Proceedings of the 34th International Conference on Machine Learning (ICML), pp. 2642–2651, 2017.
  • [20] A. Radford, L. Metz, and S. Chintala, “Unsupervised representation learning with deep convolutional generative adversarial networks,” International Conference on Learning Representations (ICLR), 2016.
  • [21] L. Tran, X. Yin, and X. Liu, “Disentangled representation learning gan for pose-invariant face recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1415–1424, 2017.
  • [22] T.-C. Wang, M.-Y. Liu, J.-Y. Zhu, G. Liu, A. Tao, J. Kautz, and B. Catanzaro, “Video-to-video synthesis,” in Advances in Neural Information Processing Systems (NeurIPS), 2018.
  • [23] Y. Chen, M. Rohrbach, Z. Yan, Y. Shuicheng, J. Feng, and Y. Kalantidis, “Graph-based global reasoning networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 433–442, 2019.
  • [24] C. Wang, M. Peng, T. Bi, and T. Chen, “Micro-attention for micro-expression recognition,” Neurocomputing, vol. 410, pp. 354–362, 2020.
  • [25] J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7132–7141, 2018.
  • [26] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, u. Kaiser, and I. Polosukhin, “Attention is all you need,” in Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, 2017.
  • [27] T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” in International Conference on Learning Representations (ICLR), 2017.
  • [28] L. Zhong, C. Bai, J. Li, T. Chen, S. Li, and Y. Liu, “A graph-structured representation with brnn for static-based facial expression recognition,” in 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), pp. 1–5, IEEE, 2019.
  • [29] S.-T. Liong, Y. Gan, D. Zheng, S.-M. Li, H.-X. Xu, H.-Z. Zhang, R.-K. Lyu, and K.-H. Liu, “Evaluation of the spatio-temporal features and gan for micro-expression recognition system,” Journal of Signal Processing Systems (JSPS), pp. 1–21, 2020.
  • [30] H.-X. Xie, L. Lo, H.-H. Shuai, and W.-H. Cheng, “Au-assisted graph attention convolutional network for micro-expression recognition,” in Proceedings of the 28th ACM International Conference on Multimedia, pp. 2871–2880, 2020.
  • [31] P. Krähenbühl and V. Koltun, “Efficient inference in fully connected crfs with gaussian edge potentials,” in Advances in neural information processing systems, pp. 109–117, 2011.
  • [32] T. Wu, Y. Lu, Y. Zhu, C. Zhang, M. Wu, Z. Ma, and G. Guo, “Ginet: Graph interaction network for scene parsing,” in European Conference on Computer Vision, pp. 34–51, Springer, 2020.
  • [33] H. Wang, Y. Song, C. Zhang, J. Yu, S. Liu, H. Peng, and W. Cai, “Single neuron segmentation using graph-based global reasoning with auxiliary skeleton loss from 3d optical microscope images,” 2021.
  • [34] X. Liang, Z. Hu, H. Zhang, L. Lin, and E. P. Xing, “Symbolic graph reasoning meets convolutions,” in Advances in Neural Information Processing Systems, pp. 1853–1863, 2018.
  • [35] L. Lo, H. Xie, H. Shuai, and W. Cheng, “MER-GCN: micro-expression recognition based on relation modeling with graph convolutional networks,” in 3rd IEEE Conference on Multimedia Information Processing and Retrieval, MIPR 2020, pp. 79–84, IEEE, 2020.
  • [36] L. Lei, J. Li, T. Chen, and S. Li, “A novel graph-tcn with a graph structured representation for micro-expression recognition,” in Proceedings of the 28th ACM International Conference on Multimedia, pp. 2237–2245, 2020.
  • [37] I. O. Ertugrul, L. A. Jeni, and J. F. Cohn, “Facscaps: Pose-independent facial action coding with capsules,” in 2018 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 2211–221109, IEEE, 2018.
  • [38] R. LaLonde and U. Bagci, “Capsules for object segmentation,” arXiv preprint arXiv:1804.04241, 2018.
  • [39] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 234–241, Springer, 2015.
  • [40] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3431–3440, 2015.
  • [41] J. Johnson, A. Alahi, and L. Fei-Fei, “Perceptual losses for real-time style transfer and super-resolution,” in European conference on computer vision (ECCV), pp. 694–711, Springer, 2016.
  • [42] Y. He, S.-J. Wang, J. Li, and M. H. Yap, “Spotting macro-and micro-expression intervals in long video sequences,” in 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG), pp. 238–244, 2020.