Deepfake Representation with Multilinear Regression
Abstract.
Generative neural network architectures such as GANs, may be used to generate synthetic instances to compensate for the lack of real data. However, they may be employed to create media that may cause social, political or economical upheaval. One emerging media is ”Deepfake”. Techniques that can discriminate between such media is indispensable. In this paper, we propose a modified multilinear (tensor) method, a combination of linear and multilinear regressions for representing fake and real data. We test our approach by representing Deepfakes with our modified multilinear (tensor) approach and perform SVM classification with encouraging results.
1. Introduction
Recent advances in Generative Adversarial Networks (GANs) and Convolutional Neural Networks (CNNs) embedded in applications like Zao111https://www.zaoapp.net/, DeepFakes web 222 https://deepfakesweb.com/, Face Swap by Microsoft333https://www.microsoft.com/en-us/garage/profiles/face-swap/, DeepFaceLab444https://awesomeopensource.com/project/iperov/DeepFaceLab etc. have led to a broad usage of AI-synthesized media a.k.a. ”Deepfake” 555The term Deepfakes has been widely used for deep learning generated media, but it is also the name of a specific manipulation technique in which face of one person is replaced by another one. To distinguish these, we denote said method by DeepFakes in the entire paper.. Other automated manipulation techniques are Face2Face, FaceSwap, NeuralTextures, and FaceShifter (roessler2019faceforensicspp, ).
Due to the potential misuse of Deepfakes e.g., fake pornography, fake news, and financial or political fraud, they have become a major public concern. Thus, different techniques have been introduced to discriminate Deepfakes from pristine videos.
Prior Deepfakes detection can be categorized as (2020_SurveyDeepFake, ) approaches that classify based on (a) physical or physiological causal factors which are not well presented in Deepfakes e.g., eye blinking (eyeblinking, ) and heart rate(heart_rate, ), or (b) artifacts in imaging factors e.g., relative head pose to the camera position(HeadPose, ), and (c) data-driven techniques that do not leverage specific cues and directly train a deep learning model on a large set of real and Deepfake videos (MesoNet, ; Xception, ).
From the first category, we can mention (HeadPose, ), where Yeng et al. propose using the inconsistencies in head poses to detect the Deepfakes. More precisely, 3D head poses cue is leveraged to estimate errors introduced by splicing process which synthesizes source face region into the target one. The eye blinking cue is anther physiological signal which is not well presented in Deepfakes and Li et al. take advantage of it for discriminating the Deepfakes (eyeblinking, ). More recently, a novel cue has been introduced that considers the heart rate measured by remote photoplethysmography (rPPG) to analyze color changes in the human skin, which is a signal for the presence of blood under the tissues (heart_rate, ).

As an example of the second category, we can refer to the work in (artifact, ) where the distinctive feature is the introduced face warping artifacts. In this work, Li et al. discuss limitation of early Deepfake generators which produce images of limited resolutions and transformation of this images leaves certain distinctive artifacts in the Deepfake videos. In addition, in (Saturation_Cues, ), McCloske et al. analyze the structure of the generator network of a GAN and show how the network’s treatment of exposure is markedly different from a real camera. They propose leveraging frequency of over-exposed pixels as a feature for this cue to discriminate GAN-generated media from camera imagery.
However, the vast majority of proposed methods for Deepfake detection fall into the third category, i.e., data-driven approaches. For instance, in (HybridLSTM_Amit, ) a hybrid Long Short Term Memory Network (LSTM) and Encoder-Decoder architecture is introduced to detect forgeries in images. In another work (Xception, ), a novel CNN network inspired by inception is introduced, where inception modules have been replaced with depth wise separable convolutions. Another example of this category is the work proposed in (MesoNet, ), where two networks are presented, both with a low number of layers to focus on the mesoscopic properties of the images. (RNN+CNN, ), (alignment, ) and (Multi_task_Learning, ) are other instances of data-driven approaches which leverage (Recurrent Neural Networks (RNNs), capsule networks and CNN networks for detection of Deepfakes. Lastly, there are works that take advantage of CNNs and RNNs simultaneously to capture both frame level and sequence level information (RNN+CNN, ; HybridLSTM_Amit, ; alignment, ).
The first two categories, which mainly leverage feature extraction and image pre-processing techniques to some extent provide interpretability for the classification result which is a key factor for explainable and trustworthy AI. For instance, the predictive model is built upon differences in the nature of pixels (Saturation_Cues, ) or an estimation of regions with high concentration of artifacts (artifact, ). However, the black box methods of the third category, while being highly accurate, do not provide any interpretation for the classification output. Thus, it is not clear if the video is classified as Deepfake due to the difference in the frame by frame movement or because of spatial artifacts or both. Moreover, there is no information about regions of interest and causes of the artifacts e.g., warping artifacts or artifacts in head adjustment, that discriminate Deepfakes from real videos.
We hypothesize that DeepFakes contain artifacts localized either in transition areas between facial images, or contain discrepancies in the overall facial appearance. We concentrate our analysis on the transition areas of the face henceforth referred to as outer facial ring, Figure 1. We segment the outer ring from a facial image that has been registered to a template based on facial landmarks detected by a pretrained model (landmark, ). The outer ring is analyzed with a modified face recognition tensor model (Vasilescu02, ; Vasilescu05, ) that computes real and fake data representations.
We employ a multilinear a.k.a. tensor framework which decomposes basis components of outer facial rings into real and fake class representations. Later on we leverage the derived representation of classes to classify the test frames using a linear SVM. Summarily, our major contributions are as follows:
-
•
Segmenting face into regions of interest: We propose Segmenting face into facial parts and leverage parts with high concentration of artifacts to distinguish Deepfakes.
-
•
Proposing a multilinear representation of Deepfakes for classification: we employ a multilinear approach to represent Deepfake and real class information and then leverage them for classification.
2. Background
In this section, we discuss the relevant tensor algebra (Delathauwer00a, ), (Delathauwer00b, ), (Vasilescu02, ; Vasilescu05, ), (KoBa09, ), (Papalexakis:2016, ), (Tensor, ). We will follow the notation of Table 1
2.1. Multilinear (tensor) framework
A data tensor is a multi-way array. In fact, when an array has three or more than three dimensions, we call it a tensor. The dimensions of a tensor are usually referred to as modes.
2.2. Singular Value Decomposition (SVD) and
Principle Components Analysis (PCA)
In linear algebra, we factorize a matrix using Singular Value Decomposition (SVD) as follows:
(1) |
where the columns of and are orthonormal and is a diagonal matrix with positive real entries know as singular values. The rank , SVD decomposition represents a matrix as following equation:
(2) |
Rewriting equation 1 in conventional linear algebra, the Principal Components Analysis (PCA) is:
(3) |
Symbol | Definition |
---|---|
, , | Tensor, Matrix, vector |
Mode-m tensor pseudo-inverse of | |
Mode- tensor matrixizing | |
Mode- product | |
Outer product |
2.3. Mode- Matrixizing a Tensor
TMode- matrixizing of tensor is defined as the matrix where the parenthetical ordering indicates that column vectors are ordered by sweeping indices of all other modes through their ranges (Vasilescu09, ). Therefore:
(4) | ||||
A -mode tensor may be metricized in three different ways by stacking first, second and third mode slices which are illustrated in Figure.2.
2.4. Mode- product of a matrix and a tensor
The mode- product (Vasilescu09, ; Carroll70, ; Delathauwer00a, ) of a tensor and matrix denoted by is a tensor of size where the entries are calculated as
(5) |
The mode- product is interchangeably denoted by matrix multiplication and tensor multiplication as follows:
(6) |

2.5. -mode SVD
We can define SVD decomposition in terms of n-mode product as follows:
(7) |
In multilinear algebra there is a generalization of SVD know as multilinear SVD (Delathauwer00a, ; Delathauwer00b, ) or -mode SVD (Vasilescu02, ; Vasilescu05, ) which decomposes an -mode tensor into the -mode product of orthonormal spaces:
(8) |
where is the core tensor that governs the interaction between the orthonormal mode matrices, . The core tensor is analogues to the singular value matrix but unlike the the core tensor is not always diagonal (KoBa09, ; Papalexakis:2016, ; Tensor, ).
The -mode SVD of a 3-mode tensor is demonstrated in Figure 3. is approximated by left singular vectors of truncated SVD decomposition of . Meanwhile, since is orthonoramal, we have and the core tensor is estimated as follows:
(9) | |||||
(10) |

3. Proposed Method
A DeepFake is a synthesizing product of two real faces. More precisely, in DeepFake generation process, face of a real person a.k.a. target is synthesized by another face a.k.a., source. This process, usually introduces some artifacts, specially around the cropping edges of source face including eyes and eyebrows Figure 1. Due to the fact that a Deepfake face is a mixture of source and target faces, Sometimes it is not distinguishable from the source and this similarity results in misclassification of the video. In this work, we propose to segment faces into parts henceforth referred to as facial inner and outer rings Figure 1. We define the outer ring as a facial part that comprises the blending boundaries that are mostly the non-facial pixels. We leverage this remaining region i.e., outer ring which has the highest concentration of introduced artifacts as a cue for Deepfakes detection. An example of this process is demonstrated in Figure. 1. This cue is very promising specially when the manipulation masks are not available.
In what follows, we discuss our proposed multilinear pipeline for detecting the Deepfakes.
3.1. Step 1: Vectorizing video frames
Vasilescu (Vasilescu09, , Appndix A) argues that in most cases, it is preferable to vectorize an image and treat it as a single observation rather than a collection of independent column/row observations. By vectorizing an image, we treat an image as a point in high dimensional pixel space and calculate all possible combinations of pixel statistics, both near and faraway statistics. On the other hand, when we consider an image as a matrix, every image column (row) is treated as an independent observation, and column (row) covariances are computed. Having this in mind, we also follow the same strategy and vectorize the frames and create a vector for each one of the video frames in the dataset.
3.2. Step 2: Finding eigenfaces of each class
Eigenfaces are eigenvectors when the images are human face. The eigenfaces are derived from the covariance matrix of the pixel distribution over the high dimensional face space. The eigenfaces represent a basis set of all faces used to construct the covariance matrix. So far, the eigenfaces have been successfully leveraged for many facial image related tasks (Turk91b, ). Leveraging eigenfaces allows for dimensionality reduction such that a smaller set of basis vectors represent the original training faces. Classification could be achieved by comparing how different faces are represented by the basis set of the corresponding class.
Based on principle component terminology, the eigenfaces are equal to basis vectors of PCA decomposition. Therefore, by staking the vectorized frames of each class, we create two separate matrices and decompose them using SVD to capture the eigenfaces of the corresponding class as follows:
(11) | |||||
(12) |
Where and are basis matrices and and are the normalized coefficient matrices of the corresponding classes.
3.3. Step 3: Leveraging tensor framework to decompose eigenfaces into underlying factors
As seen in previously mentioned tensor is an effective framework for decomposing a set of observation into underlying factors. After reducing the dimentionality of observations using eigenface representation of the classes, we propose leveraging a three-mode tensor where the first mode i.e., measurement mode represents the pixels of an eigenface, the second mode corresponds to the eigenfaces and the third mode is the class mode i.e., DeepFake vs. real. We propose using an -mode SVD which as we discussed earlier decomposes a tensor into orthonormal matrices (), and a core tensor which governs the interaction between these spaces. Since the first mode is the measurement mode, We only calculate the -mode SVD of the tensor by flattening the second and the third modes as follows:
(13) | |||||
(14) |
where the comprises underlying vector representation of original and fake classes.Moreover, the core tensor is the signature of this dataset and shows interactions of orthonormal subspaces. Later on, we leverage this signature to project the test frames into the subspaces we derive here.
3.4. Step 4: Embedding the class representations in a higher three dimensional space
Applying -mode SVD results in a mode matrix that spans the class representations. We embed the vector class representations into a higher dimensional space to increase the class separability of the test data. We embed the row vectors of into , setting the third coordinate of the real and fake class to and respectively, and normalizing the vector length to .
3.5. Step 5: Multilinear projection of an incoming frame into the orthonormal vector spaces
As mentioned above, the core tensor of each decomposition is the signature of the decomposed space which governs the interaction of constituent factors. we leverage the core tensor and perform a multilinear projection of the incoming frame into the subspaces we derived in the previous step. Let say we have the vectorized frame . If is supposed to be in the same subspaces we derived, then
(15) |
where the vectors and are the coefficient vector representations of a video frame in the orthonormal subspaces that are governed by the extended core tensor . The goal is to find out weather the class coefficient vector is more similar to the vector representation of real class or Deepfake class. To this end, we estimate representation vector by employing the multilinear projection algorithm (Vasilescu11, ; Multilinear_Projection2007, ) that decomposes a vectorized observation, into a set of latent vector representation, that corresponds to the constituent factors of data formation. The basic multilinear projection is the -mode SVD/CP decomposition of which can be expressed mathematically as
where is mode-1 pseudo-inverse of that in matrix notation is expressed as , and are estimates of vectors and from eq.(15), respectively.
3.6. Step 6: Classifying an incoming frame
Up to this step, we have the vector representation of each classes in addition to class coefficients of the incoming frame. We use a linear Support Vector Machine (SVM) and estimate the decision boundaries using validation frames and then leverage the defined boundaries for classification of test frames. An overview of the proposed approach is demonstrated in Algorithm 1.
3.7. Dimensionality reduction in step 3
As mentioned earlier, factor matrix comprises underlying structures of basis vector continent. Despite the fact that we construct our predictive model by approximating discriminating regions, still there are many shared components which getting rid of them make the model more distinguishable. Since we are interested in noisy regions i.e., artifacts, we propose truncating components of the core tensor which correspond to top values of and keeping lower value components as representatives of noisy parts. In the next section, we will show how this truncation boosts the classification performance of the proposed framework. An example of truncating components corresponding to the second mode of a 3-mode tensor is depicted in Figure. 3.
were centered by subtracting the mean of the real training data,
-
(1)
Preprocessing and data tensor organization:
-
(2)
Training data decomposition:
-
(3)
Embed the class representations in the higher three dimensional space and set the third coordinate of the real and fake class to and respectively. Hence, now has dimensionality . Normalize the rows of to have length .
-
(4)
Computer the extended core
(16) -
(5)
Centering: validation and test data is centered by subtracting the mean of the real training data.
-
(6)
Test data decomposition of a centered :
-
(7)
Finding linear SVM decision boundaries using validation set
-
(8)
classifying all test set

4. Experimental Evaluation
In this section, we first introduce the dataset and benchmark on this dataset and then we discuss the implementation details and the experimental evaluation.
4.1. Dataset description
One of the most popular and widely used databases for image or video forgeries detection is FaceForensics++666https://github.com/ondyari/FaceForensics which first was introduced in 2018 (roessler2018faceforensics, ). FaceForensics++ comprises more than 500,000 frames from 1000 youtube videos that contain mostly frontal faces (roessler2019faceforensicspp, ). This dataset also includes 1000 videos which are the manipulated version of the original onesand have been manipulated by four automated face manipulation methods: Deepfakes, Face2Face, FaceSwap and NeuralTextures. All original and manipulated videos have constant frame rate of 30 fps and have been compressed lossless with H.264. Moreover, the videos are split up into train set of size 720, validation set of size 140 and test of 140 videos. binary classification scenario on this dataset. A summarized benchmark of existing techniques on videos manipulated by DeepFakes method is demonstrated in table 2. The state-of-the-art benchmark on FaceForensics++ is available in GitHub777http://kaldir.vc.in.tum.de/faceforensics-benchmark/. In this work, we experiment on images manipulated by DeepFake technique.
Method | Accuracy |
---|---|
ZAntiFakeBio | 1.000 |
Leo | 1.000 |
Aquarius | 1.000 |
RobustForensics | 0.991 |
NoSenseAtAll | 0.982 |
PredictFake | 0.973 |
Cancer | 0.964 |
Balance | 0.918 |
unet+res | 0.882 |
HRC | 0.827 |
GAEL-Net | 0.718 |
.
4.2. Implementation
Our work was implemented in MATLAB partially using Tensor Toolbox version 2.6. (TTB_Sparse, ; TTB_Software, ). Since all videos have constant frame rate 30 fps, we extracted up to 7 frames for each video by snapping almost one frame per each 30 seconds using OpenCV library in Python. Moreover, for detecting facial landmarks, we used pretrained dlib face detector888http://dlib.net/face_landmark_detection.py.html which is created using the classic Histogram of Oriented Gradients (HOG) feature combined with a linear classifier, an image pyramid, and sliding window detection scheme (landmark, ). For the second step, we calculated the SVD rank r where the r is equal to ”” for all experiments. The intuition behind this estimation is to have an individual component for each frame. Moreover, in contrast to many deep learning approaches for Deepfake detection, our approach does not require GPU base configuration and both train and test steps can be executed on an ordinary CPU based configuration. The description of the CPU based configuration we experimented on is as follows: Intel(R) Core (TM) i5-8600K CPU @3.60GHz,CentOS Linux 7 (Core) operating system and 40GB RAM memory.
4.3. Evaluation
4.3.1. Classification performance
Classification performance of our proposed multilinear framework when we keep all of the components as well as when we truncate different ranges of components, is illustrated in Figure. 4. In this Figure, TN, TP, and ACC. denote true negative, true positive, and accuracy respectively.
As demonstrated, truncating top and bottom components, significantly improves the classification accuracy. In this work, we aim to find discriminating representations for outer ring of real vs Deepfake videos introduced by synthesizing artifacts. Thus, we hypothesize the noisy components i.e., components with insignificant values may represent those artifacts. So, by truncating the top components, we avoid high level facial structures and only keep those that correspond to what we aim to capture i.e., artifacts, for the classification. Moreover, the last components are the most insignificant ones that might be introduced by noises other than synthesizing artifacts. Anyhow, keeping components in range results in around accuracy.
4.3.2. Effects of truncation on class representations
To clarify the efficacy of truncation, we depict the PCA coefficients of column vectors of for test frames before and after applying truncation. The distribution of the PCA coefficients is demonstrated in Figure.5. As shown, truncating the undiscriminating components, makes coefficients of each class more similar and as a result put them closer to each other. Specially in case of samples that are located in outer parts of the semicircle i.e., outliers. In other words, the representations after truncation are more linearly separable than those before applying the truncation.
5. Conclusion and Future Work
In this work, we leverage the region that we hypothesize has highest concentration of artifacts, the face outer ring, for classification of Deepfakes using our proposed multilinear framework. Our preliminary results show that using only the outer facial ring we achieve 82% accuracy. In future work, we will learn class representations by subdividing an image into parts (Vasilescu92, ) and treating them as either items in a part-based hierarchy or as items in a ”bag of parts” whose representations may learned bottom-up (Vasilescu20, ). Another direction for future work, is to use binary masks released by (roessler2019faceforensicspp, ). The binary mask can be leveraged for precise segmentation of the frames into regions of interest.
6. Acknowledgements
The authors would like to thank Ghazal Mazaheri and Amit Roy-Chowdhury for initial help with the dataset. Research was supported by the National Science Foundation CDS&E Grant no. OAC- and a UCR Regents Faculty Fellowship. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the funding parties.


References
- (1) Darius Afchar, Vincent Nozick, Junichi Yamagishi, and I. Echizen. Mesonet: a compact facial video forgery detection network. 09 2018.
- (2) Brett W. Bader and Tamara G. Kolda. Efficient MATLAB computations with sparse and factored tensors. SIAM Journal on Scientific Computing, 30(1):205–231, December 2007.
- (3) Brett W. Bader, Tamara G. Kolda, et al. Matlab tensor toolbox version 2.6. Available online, February 2015.
- (4) Md Jawadul Bappy, Cody Simons, Lakshmanan Nataraj, B. Manjunath, and Amit Roy-Chowdhury. Hybrid lstm and encoder-decoder architecture for detection of image forgeries. IEEE Transactions on Image Processing, PP, 01 2019.
- (5) J. D. Carroll and J. J. Chang. Analysis of individual differences in multidimensional scaling via an N-way generalization of ‘Eckart-Young’ decomposition. Psychometrika, 35:283–319, 1970.
- (6) François Chollet. Xception: Deep learning with depthwise separable convolutions. pages 1251–1258, 2017.
- (7) L. de Lathauwer, B. de Moor, and J. Vandewalle. A multilinear singular value decomposition. SIAM J. of Matrix Analysis and Applications, 21(4):1253–78, 2000.
- (8) L. de Lathauwer, B. de Moor, and J. Vandewalle. On the best rank-1 and rank-() approximation of higher-order tensors. SIAM J. of Matrix Analysis and Applications, 21(4):1324–42, 2000.
- (9) David Guera and Edward Delp. Deepfake video detection using recurrent neural networks. pages 1–6, 11 2018.
- (10) Javier Hernandez-Ortega, Ruben Tolosana, Julian Fierrez, and Aythami Morales. Deepfakeson-phys: Deepfakes detection based on heart rate estimation, 2020.
- (11) Vahid Kazemi and Josephine Sullivan. One millisecond face alignment with an ensemble of regression trees. In 2014 IEEE Conference on Computer Vision and Pattern Recognition, pages 1867–1874, 2014.
- (12) Tamara G. Kolda and Brett W. Bader. Tensor decompositions and applications. SIAM Review, 51(3):455–500, September 2009.
- (13) Yuezun Li, Ming-Ching Chang, Hany Farid, and Siwei Lyu. In ictu oculi: Exposing ai generated fake face videos by detecting eye blinking. 06 2018.
- (14) Yuezun Li and Siwei Lyu. Exposing deepfake videos by detecting face warping artifacts. CoRR, abs/1811.00656, 2018.
- (15) S. McCloskey and M. Albright. Detecting gan-generated imagery using saturation cues. In 2019 IEEE International Conference on Image Processing (ICIP), pages 4584–4588, 2019.
- (16) Huy Nguyen, Fuming Fang, Junichi Yamagishi, and I. Echizen. Multi-task learning for detecting and segmenting manipulated facial images and videos. 06 2019.
- (17) Huy Nguyen, Junichi Yamagishi, and I. Echizen. Use of a capsule network to detect fake images and videos. 10 2019.
- (18) Evangelos E. Papalexakis, Christos Faloutsos, and Nicholas D. Sidiropoulos. Tensors for data mining and data fusion: Models, applications, and scalable algorithms. ACM Trans. Intell. Syst. Technol., 8(2):16:1–16:44, Oct. 2016.
- (19) Andreas Rössler, Davide Cozzolino, Luisa Verdoliva, Christian Riess, Justus Thies, and Matthias Nießner. Faceforensics: A large-scale video dataset for forgery detection in human faces. arXiv, 2018.
- (20) Andreas Rössler, Davide Cozzolino, Luisa Verdoliva, Christian Riess, Justus Thies, and Matthias Nießner. FaceForensics++: Learning to detect manipulated facial images. In International Conference on Computer Vision (ICCV), 2019.
- (21) N.D. Sidiropoulos, Lieven De Lathauwer, Xiao Fu, Kejun Huang, Evangelos Papalexakis, and Christos Faloutsos. Tensor decomposition for signal processing and machine learning. IEEE Transactions on Signal Processing, PP, 07 2016.
- (22) Ruben Tolosana, Ruben Vera-Rodriguez, Julian Fierrez, Aythami Morales, and Javier Ortega-Garcia. Deepfakes and beyond: A survey of face manipulation and fake detection. arXiv preprint arXiv:2001.00179, 2020.
- (23) Mathew A. Turk and Alex P. Pentland. Eigenfaces for recognition. Journal of Cognitive Neuroscience, 3(1):71–86, 1991.
- (24) M. Vasilescu and D. Terzopoulos. Adaptive meshes and shells: Irregular triangulation, discontinuities, and hierarchical subdivision. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition (CVPR’92), page 829–832, Champaign, IL, Jun 1992.
- (25) M. Alex O. Vasilescu. A Multilinear (Tensor) Algebraic Framework for Computer Graphics, Computer Vision, and Machine Learning. PhD thesis, University of Toronto, 2009.
- (26) M. A. O. Vasilescu. Multilinear projection for face recognition via canonical decomposition. In Proc. IEEE Inter. Conf. on Automatic Face Gesture Recognition (FG 2011), pages 476–483, Mar 2011.
- (27) M. Alex O. Vasilescu, Eric Kim, and Xiao S. Zeng. Causalx: Causal explanations and block multilinear factor analysis. In 2020 25th International Conference of Pattern Recognition (ICPR 2020), pages 10736–10743, Jan 2021.
- (28) M. A. O. Vasilescu and D. Terzopoulos. Multilinear analysis of image ensembles: TensorFaces. In Proc. European Conf. on Computer Vision (ECCV 2002), pages 447–460, Copenhagen, Denmark, May 2002.
- (29) M. A. O. Vasilescu and D. Terzopoulos. Multilinear independent components analysis. In Proc. IEEE Conf. on Computer Vision and Pattern Recognition, volume I, pages 547–553, San Diego, CA, 2005.
- (30) M. Alex O. Vasilescu and Demetri Terzopoulos. Multilinear projection for appearance-based recognition in the tensor framework. In IEEE 11th International Conference on Computer Vision, ICCV 2007, Rio de Janeiro, Brazil, October 14-20, 2007, pages 1–8. IEEE Computer Society, 2007.
- (31) X. Yang, Y. Li, and S. Lyu. Exposing deep fakes using inconsistent head poses. pages 8261–8265, 2019.