\ul
Style-Preserving Lip Sync via Audio-Aware Style Reference
Abstract
Audio-driven lip sync has recently drawn significant attention due to its widespread application in the multimedia domain. Individuals exhibit distinct lip shapes when speaking the same utterance, attributed to the unique speaking styles of individuals, posing a notable challenge for audio-driven lip sync. Earlier methods for such task often bypassed the modeling of personalized speaking styles, resulting in sub-optimal lip sync conforming to the general styles. Recent lip sync techniques attempt to guide the lip sync for arbitrary audio by aggregating information from a style reference video, yet they can not preserve the speaking styles well due to their inaccuracy in style aggregation. This work proposes an innovative audio-aware style reference scheme that effectively leverages the relationships between input audio and reference audio from style reference video to address the style-preserving audio-driven lip sync. Specifically, we first develop an advanced Transformer-based model adept at predicting lip motion corresponding to the input audio, augmented by the style information aggregated through cross-attention layers from style reference video. Afterwards, to better render the lip motion into realistic talking face video, we devise a conditional latent diffusion model, integrating lip motion through modulated convolutional layers and fusing reference facial images via spatial cross-attention layers. Extensive experiments validate the efficacy of the proposed approach in achieving precise lip sync, preserving speaking styles, and generating high-fidelity, realistic talking face videos.
Index Terms:
Lip Sync, Talking Face Generation, Style-Preserving, Audio-Aware ReferenceI Introduction
Audio-driven lip sync, also referred to as talking face generation, aims to generate a talking video having lip motion synchronized with the input audio. This technology has garnered significant attention in the multimedia domain in recent years, spanning applications such as video translation [1], visual dubbing [2], virtual avatars [3], and digital humans [4]. Traditionally, achieving accurate lip sync has primarily involved training subject-specific models with data collected from specific individuals. However, these techniques are hampered by the significant costs associated with subject-specific training and data collection processes. Consequently, we focus on addressing the task of audio-driven lip sync through a subject-generic strategy [5, 6, 7, 8, 9, 10], which, by training once on a large-scale dataset, can be generalized to any individual without the need for further training or fine-tuning.

In general, when individuals speak the same utterance, variations in their speaking styles lead to differences in lip shapes [11, 6, 5, 12, 13, 14, 15], posing a challenge to accurate lip sync. However, most previous subject-generic approaches, such as [7, 8, 9, 10, 2, 16, 1], have overlooked modeling the speaking style of individuals, as shown in Figure 1(a). These approaches merely predict lip motion from audio without incorporating any reference information regarding the speaking style of subject. As a result, these subject-generic approaches typically only achieve lip sync consistent with the general speaking styles learned from audio-visual datasets. There are still discrepancies between the lip sync conforming to general speaking styles and those customized to individual speaking styles.
To alleviate this issue, recent methods [6, 5, 11, 12, 13, 17] have attempted to incorporate a short style reference video of the individual into the lip sync model as illustrated in Figure 1(b). They learn a video-level style code from the reference video to guide the lip sync for arbitrary audio. However, predicting lip shapes for different audio requires referencing different lip shapes from the style reference video. Therefore, the compact and static video-level style codes learned by these methods are still insufficient to achieve satisfactory lip sync.
Inspired by the observation from [18] that similar pronunciations produce similar lip shapes, this paper posits that within a style reference video, if a lip shape corresponds to the reference audio that is closer to the input audio, it should contribute more significantly to the style coding. Thus, we approach Audio-driven Lip Sync from a novel perspective and propose an innovative audio-aware style reference scheme. As shown in Figure 1(c), our method leverages the audio signals in the style reference video to aggregate distinct style codes for different input audios, facilitating Style-Preserving Lip Sync.
Specifically, we propose a novel Transformer-based architecture designed to predict lip motions from audio inputs, influenced by a style reference video. The process commences with the extraction of lip shapes from a style reference video utilizing the 3D Morphable Model (3DMM) [19]. Subsequently, these lip shapes and their associated audio signals are encoded via dual Transformer encoders, yielding respective representations for reference lip and audio features. Input audio information is then integrated into a Transformer decoder to generate hidden features, which are further processed by a cross-attention layer within the decoder to aggregate style information from the style reference video. Herein, the reference audio and lip features serve as keys and values, respectively, while the hidden features derived from the input audio act as queries. Ultimately, the Transformer decoder outputs mouth-related expression parameters representing lip motion. To render realistic talking face videos from lip motion, this research devises a conditional latent diffusion model [20] tailored for producing lip-synced faces. Specifically, the integration of mouth-related expression parameters into the latent diffusion model is achieved through modulated convolutional layers. This approach ensures that the synthesized lip shapes are aligned with those represented by mouth-related expression parameters. Additionally, the incorporation of reference facial image into the diffusion model is facilitated via spatial cross-attention layers, which enhance the model’s capacity to capture subject-specific appearance details, thereby improving the fidelity of the generated talking faces.
The primary contributions of this study are summarized as follows.
-
•
We propose a novel audio-aware style reference scheme that leverages the intrinsic connections between input and reference audios to effectively aggregate speaking style information, thereby achieving style-preserving lip synchronization.
-
•
We have devised an advanced Transformer-based lip sync model that integrates speaking style information from a style reference video to predict lip motions. Additionally, we introduced a novel conditional latent diffusion model to translate the predicted lip motions into lip-synced and realistic videos.
-
•
Extensive experiments demonstrate that our proposed method achieves style-preserving lip synchronization, which effectively preserves the speaking style of the individual, resulting in realistic and high-fidelity talking face videos.
II RELATED WORK
II-A Audio-Driven Lip Sync
With the increasing application in multimedia, audio-driven lip sync has attracted widespread attention in recent years. According to the training paradigm, methods for audio-driven lip sync can be categorized into subject-specific and subject-generic. Subject-specific methods[18, 21, 22, 23, 24, 25] need to collect data of the target subject for subject-specific training. Although they can achieve lip sync conforming to the individual’s speaking style, they can not be generalized to unseen subjects without further training, thus having limited application in the real world. Related to ours, AE-NeRF [18] observes that similar pronunciations have similar lip shapes. However, instead of aggregating the reference lip shapes to model the speaking style, it aggregates the reference image features to build a robust prior for Neural radiance field (NeRF) [26] and can not generalize across different subjects without further training.
In contrast, subject-generic approaches for lip sync, such as [5, 6, 7, 8, 9, 10, 16, 1, 2], have garnered more attention for their ability to infer lip sync for unseen subjects without further training. However, most subject-generic methods [7, 8, 9, 10, 2, 16, 1] neglect to model the speaking styles of individuals in the absence of reference information regarding individual speaking style. For example, Wav2Lip [7] pre-trains a lip synchronization discriminator on a large-scale audio-visual dataset to guide the training of a lip sync generator. Yet, it can only achieve lip sync matching the general speaking styles learned from the audio-visual dataset. Similarly, IP-LAP [8] proposes a two-stage talking face generation method based on landmark representation, using a transformer-based model to predict lip landmarks from input audio. However, due to the lack of reference information on individual speaking style, it can not generate lip motions that preserve the speaking style well.
In this paper, we propose a method that further improves subject-generic methods, achieving more accurate lip sync that better preserves the individual’s speaking style.
II-B Style-Guided Lip Sync
Recent advancements in subject-generic methods, e.g. [6, 5, 11, 12, 13, 17], have made efforts to model the speaking styles of individuals through the integration of style reference videos. Among these, StyleAvatar [13] constructs a video-level style code from a style reference video based on the standard deviation of facial movements. Then, it proposes a latent-style-fusion model to synthesize stylized talking faces for arbitrary audio, imitating the talking style from the style code. Recently, StyleTalk [6] introduces a framework for generating stylized one-shot talking heads. This framework encodes the expression parameters from all frames within a style reference video into style vectors. These vectors are then aggregated into a single video-level style code through a self-attention pooling layer. Specifically, this layer calculates the attention weight for each frame by utilizing a feed-forward network, subsequently aggregating the style vectors weighted by their respective attention weights into a video-level style code. The derived style code is then used to guide the lip sync for arbitrary input audio.
Following StyleTalk [6], some later works [5, 11, 12, 17] adopt a similar style aggregation method to obtain a compact video-level style code. Despite their efforts, a common limitation persists across these methods: the inability to well preserve the speaking style of the individual during the task of lip sync. The main reason for this challenge is that, to predict the lip shape for different input audio, the network needs to reference different lip shapes from the style reference video. In response to this challenge, our study proposes a novel audio-aware style reference scheme to preserve the speaking style better for improved lip sync.
II-C Diffusion Models
Diffusion models [27] have emerged as a significant advancement in generative tasks. Compared to generative adversarial networks (GANs) [28], diffusion models exhibit enhanced training stability and alleviate the mode collapse issue, achieving superior generation quality. Particularly, latent diffusion models (LDM) [20] apply the diffusion model in the latent space of a pre-trained autoencoder to reduce the computational cost while retaining the generation quality. A further innovation is proposed in Denoising Diffusion Implicit Models (DDIM) [29], which introduce a class of non-Markovian diffusion processes to accelerate the inference process of diffusion models.
Among the applications of diffusion models, DiffTalk [10] designs a conditional latent diffusion model for lip sync tasks but overlooks modeling the speaking styles of individuals. It incorporates input audio through cross-attention layers to facilitate lip sync. Moreover, it concatenates the reference image’s latent with the noisy latent to provide subject appearance information for the latent diffusion model. In this work, we propose an advanced conditional latent diffusion model to generate lip-synced and realistic talking face videos, conditioned on the lip motion and a reference facial image.
III METHODOLOGY

III-A Preliminaries
III-A1 Problem Definition
The task of audio-driven lip sync is to generate lip-synced videos given a segment of audio, a video, and a style reference video as inputs. In this task, the lower-half face of the input video is masked, and then the masked region is inpainted with lip-synced content by the proposed method.
III-A2 3D Morphable Model (3DMM)
According to [19], the 3DMM mainly serves as a comprehensive three-dimensional facial model that represents facial shape using a fixed number of points. Specifically, the facial shape can be depicted using a facial mesh with vertices, and such a facial mesh can be parameterized via Principal Component Analysis (PCA) as follows,
(1) |
Here, denotes the average facial shape, while and represent the PCA bases for identity and expression, respectively. The parameters and control facial identity and expression, respectively. Head rotation is expressed using another set of parameters, .
Typically, off-the-shelf 3D face reconstruction models enable the extraction of the parameters from any real-world facial image. It is important to note that of the 64 expression parameters, only 13 are significantly related to mouth motions, as illustrated by [6, 5]. Therefore, in this work, we utilize these 13 mouth-related expression parameters to represent lip motion.
III-A3 Method Overview
An overview of our method is illustrated in Figure 2. In this study, we present a novel two-stage framework called StyleSync, which introduces an innovative audio-aware style reference scheme for style-preserving lip sync. Specifically, we first propose a module named Reference-guided Lip Motion Prediction, aimed at predicting lip shapes from input audio signals guided by a style reference video. This is achieved through an innovative Transformer-based model that aggregates speaking style information from the style reference video. Then, StyleSync introduces another module called Realistic Face Rendering, which utilizes a conditional latent diffusion model to produce realistic talking face videos.
III-B Reference-guided Lip Motion Prediction
To accomplish the task of predicting lip motions, we initiate the process with reference audio and lip shape encoding. This step aims to acquire reference audio and lip features from the style reference video. Subsequently, we employ the cross-attention layers in transformer decoder to aggregate the speaking style information from these reference audio and lip features. The aggregated style is then employed to guide the lip motion prediction.
III-B1 Reference Audio and Lip Shape Encoding
A style reference video of an individual contains the lip shapes produced when that individual pronounces some utterances, reflecting the speaking style. Initially, we employ DeepSpeech [30] to extract the per-frame DeepSpeech features of the style reference video, denoted as , where represents the length of the style reference video. These DeepSpeech features are further processed by a 1D temporal convolutional module and fed into a Transformer encoder, yielding reference audio features .
To obtain the corresponding lip shape representation from the style reference video, we utilize an off-the-shelf 3D face reconstruction model [31] to extract 3DMM expression parameters from video frames. We then utilize the mouth-related expression parameters to represent the reference lip shapes, denoted as . These mouth-related expression parameters are inputted into another Transformer encoder, generating reference lip features .
III-B2 Audio-Aware Style Aggregation in Transformer
Once obtaining the speaking style information from the style reference video, represented by reference audio and lip features, we utilize a Transformer decoder [32] for generating mouth-related expression parameters corresponding to the input audio. As illustrated in Figure 2, the Transformer decoder consists of Transformer blocks, each comprising a stack of self-attention, cross-attention, and feed-forward layers.
To capture temporal coherence, the Transformer decoder takes audio features from a temporal window as input, where represents the window size and denotes the current time step. Guided by the reference audio and lip features, it predicts the lip motion at the middle time step . To achieve this goal, we first encode the DeepSpeech features of the temporal window into audio tokens using a 1D convolutional module, denoted as . Subsequently, these audio tokens are inputted into the self-attention layer of the transformer decoder to generate hidden features .
To achieve style aggregation, previous methods, such as [6, 5, 33], utilize a self-attention pooling layer on reference lip features to aggregate the style information. Unlike these previous approaches, this work proposes a novel audio-aware style aggregation scheme using cross-attention layers.
Specifically, we begin by treating the obtained reference audio features as keys , and the corresponding reference lip features as values , where represents the dimension of the key. These keys and values are inputted into the cross-attention mechanism. Simultaneously, the hidden features of temporal window act as queries in the cross-attention mechanism, denoted as . Therefore, the style reference information can be aggregated through the cross-attention layer, formulated as follows,
(2) |
where denotes the aggregated styles for the temporal window at time step .
It is noteworthy that, when predicting lip motions for different audio inputs, the attention weights vary based on the similarity between the reference audios and the audios of the temporal window. The closer the reference audio is to the audio within the temporal window, the more its corresponding reference lip shape contributes to aggregating the style. This differs from prior works [6, 5, 33, 11], which aggregate a video-level style remaining static for predicting lip motion for arbitrary audio.
The aggregated styles are then added to the hidden features in a residual manner, which are further fed into the feed-forward layer. As a result, the Transformer decoder outputs vectors, among which the middle vector, corresponding to the ()-th position, is inputted into an MLP layer to generate mouth-related expression parameters, denoted by , for the time step .
III-B3 Training Objectives
During the stage of lip motion prediction, the optimization involves L1 reconstruction for both mouth-related expression parameters and lip vertices, as detailed below.
As the proposed StyleSync predicts lip motion for each frame individually, ensuring temporal consistency in Lip Motion Prediction is crucial. To achieve this, we follow the common practices of previous methods[34, 6, 5] by first generating the lip motions of successive frames for each video during training. Subsequently, we utilize the following L1 reconstruction loss for mouth-related expression parameters:
(3) |
where represents the ground truth mouth-related expression parameters for the -th frame.
Additionally, since mouth-related expression parameters may contain some information unrelated to lip shape, we also impose an L1 reconstruction loss on lip vertices. Specifically, the predicted mouth-related expression parameters can be combined with other parameters and then converted into the face mesh vertices using Equation (1). We denote the lip vertices in the facial mesh as , and the reconstruction loss on lip vertices can be expressed as:
(4) |
where represents the ground truth lip vertices detected by face alignment tools [35].
Overall, the optimization objective of this stage can be summarized as follows,
(5) |
where is a scalar for loss item balance, empirically set to 300.
III-C Realistic Face Rendering via Diffusion
To construct the conditional face rendering model tailored for producing lip-synced faces, we first introduce the latent diffusion model (LDM) proposed by [20]. LDM originates from the diffusion models (DDPM) [27], which employ a diffusion process to gradually introduce noise into the real data and train a denoising network to reverse the diffusion process. Once trained, the diffusion models can synthesize data by iteratively denoising a normally distributed variable. In comparison to DDPM, LDM [20] performs diffusion and denoising processes in the latent space of a pre-trained autoencoder composed of an encoder and a decoder , aiming to reduce the computational cost. Specifically, given an input image , the image is first encoded into the latent, i.e. , and then diffused to with time step . Afterwards, LDM trains a U-Net-based [36] denoising network to predict the noise added to the image latent , with denoting its learnable parameters. Thus, the training objective of LDM can be formulated as follows,
(6) |
where is the ground-truth noise added to the image latent and is uniformly sampled from . During inference, LDM progressively denoise a normally distributed variable until it reaches a clean latent , which can be decoded by to synthesize image.
In this work, since we focus on synchronizing the lip shape in the lower-half face, we conduct both the diffusion and denoising processes in the lower half of the encoded latent as illustrated in Figure 2. Simultaneously, to provide additional context, the upper half of the encoded latent alongside the lower-half portion are integrated into the denoising U-Net. During training, we encode the ground-truth face into which is then diffused to by the noise. During inference, we mask the lower half of the input face and encode it into , of which the lower half is diffused to obtain the initial .
To generate lip-synced talking video, we need to condition the latent diffusion model on the predicted lip motion obtained as detailed in Section III-B. Besides, it is also important to involve the reference facial image condition into the diffusion model, aiming to preserve more appearance details of the target subject. Next, we will introduce the conditioning mechanisms used to address these challenges. These include integrating mouth-related expression parameters into LDM for lip motion alignment via modulated convolution, and incorporating reference facial image into LDM for high-fidelity generation via spatial cross-attention.
III-C1 Aligning Lip Motion via Modulated Convolution
In the lip motion prediction stage, we have predicted the mouth-related expression parameters representing the lip motion synchronized to the input audio. To ensure that the synthesized facial image has lip shape aligned with the one reflected by the mouth-related expression parameters, we integrate the lip motion condition into the latent diffusion model via modulated convolutional layers, drawing inspiration from StyleGAN2 [37].
Specifically, as shown in the Figure 2, the mouth-related expression parameters are first combined with the identity and rotation parameters extracted from the input video frame, aiming to preserve the subject identity and head rotation. The combined parameters are then encoded to a latent code by a 1D convolutional module. The latent code is employed to modulate the convolution weights in the convolutional layers of denoising U-Net. Formally, we denote the original convolution weights as . The latent code is mapped to a scale set by an MLP layer, where is the scale coefficient corresponding to the -th input feature map. Then, we compute the weight of modulated convolution as follow,
(7) |
where indicates the convolution weight for the -th output features map, indicates the spatial footprint of the convolution, and is a small constant for avoiding numerical issues.
Finally, we perform the convolution operation with the modulated convolution weight to produce the output features map.
III-C2 High-fidelity Generation with Spatial Attention
To ensure the generated face has high-fidelity appearance of the subject, a reference facial image should be provided as a condition for the denoising network. As shown in the Figure 2, the reference image is initially encoded into the latent space as . To preserve more appearance details from the reference image, we devise an appearance encoder consisting of residual convolution blocks. The appearance encoder converts the reference latent into the multi-scale reference features. Given the efficacy of preserving appearance details from the reference image proven by [38, 39], we then integrate the resulting reference features into the denoising network via spatial cross-attention. Specifically, the reference features serve as the keys and values for spatial cross-attention layers, while the hidden features after modulated convolutional layer in the denoising network serve as the queries. Finally, the output of the spatial cross-attention layer are added to the hidden features in a residual manner.
III-C3 Training and Inference
For optimization, we train the diffusion rendering model using the objective as defined in Equation 6. The latent diffusion model generates the facial image of each frame individually, thus enabling the generation of videos with arbitrary length. As well, for achieving temporal consistency, we adopt the batched sequential training strategy [34] to generate five sequential images for each video during training. For model inference during testing, the denoising processes for all frames share the same initial random noise. Note that we randomly select the reference facial image from the same video during training, but alternatively, choose any facial image reflecting the subject appearance as reference during inference.
IV EXPERIMENTS
IV-A Experimental Setups
IV-A1 Datasets
In this study, we conduct experiments on two widely utilized audio-visual datasets: VoxCeleb [40] and HDTF [14]. The VoxCeleb dataset comprises over 100,000 utterances from 1,251 celebrities, sourced from videos uploaded to YouTube. It showcases a broad diversity in terms of ethnicities, accents, and ages, and is a standard benchmark commonly used in prior work. On the other hand, the HDTF dataset is a high-resolution audio-visual collection consisting of approximately 362 unique videos spanning over 15.8 hours, with resolutions of 720P or 1080P. To facilitate a fair comparison, all comparison methods, along with the proposed StyleSync, are trained on VoxCeleb and tested on both datasets.
IV-A2 Implementation Details
The videos are resampled to 25 fps, and the facial images cropped from the video frames are resized to 256256 resolution. The proposed method is composed of two stages, each trained separately. For Reference-guided Lip Motion Prediction, the style reference video is extracted from the talking clip of the same individual at different time steps. Importantly, there is no overlap between the style reference video and the ground-truth video during testing. The length of the style reference video is set to 256 following previous methods [6]. The transformer decoder consists of blocks, with the temporal window size set to 5. The query dimension is 256, and we utilize the multi-head attention layer with 8 heads. In the stage of Realistic Face Rendering, the number of time steps for diffusion, denoted as , is 1000. The latent space of the pre-trained autoencoder has a spatial resolution of 6464. During inference, we employ the DDIM [29] sampler with 200 steps to accelerate the generation process. The reference facial image is randomly selected from the input video, and all comparison methods use the same reference facial image for a fair evaluation. After obtaining the generated face, we blend it with the background following the previous method[8], producing final talking face videos.
We train all modules using the Adam optimizer [41]. The lip motion prediction network is trained with a batch size of 256 and a learning rate of 2e-4 on 4 NVIDIA 3090 GPUs. In our latent diffusion framework, the autoencoder is initialized with the pre-trained weight from latent-diffusion [42] and is frozen during training. The latent diffusion model is trained with a batch size of 12 and a learning rate of 4e-5 on 4 NVIDIA 3090 GPUs.
IV-A3 Evaluation Metrics
In the task of audio-driven lip sync, the evaluation mainly focuses on the accuracy of lip sync and the visual quality of the generated facial video. For the assessment of speaking style preservation in lip sync, this study utilizes the widely adopted metric called Lip Landmarks Distance (LipLMD) [43], which quantifies the discrepancies between the generated lip shapes and the ground-truth lip shapes corresponding to the input audio. A lower score of LipLMD indicates that the generated lip shapes are closer to the ground-truth lip shapes reflecting the individual’s speaking style, signifying a more accurate preservation of the speaking style in the generated lip sync. Furthermore, SyncScore [44] is another commonly used metric for evaluating the synchronization of lip shape with the input audio. The SyncScore is computed based on the distance between the audio feature and visual feature in SyncNet [44]. However, since SyncNet does not model the speaking styles of the individuals, SyncScore primarily assesses whether the generated lip shape matches the lip shape conforming to the general speaking style learned from a large-scale audio-visual dataset. As previously mentioned, there are still discrepancies between the lip shapes conforming to the general speaking styles and those customized to the individual’s speaking style. Therefore, it is worth noting that the SyncScore is insufficient to assess the speaking style preservation. A better SyncScore only indicates that the distance between audio and visual features in SyncNet is closer, which does not necessarily mean that the generated lip shapes are more accurate and preserve the speaking style better. Despite reporting results for this metric, we recommend LipLMD as the preferred indicator for evaluating lip sync quality since our study focuses on the task of style-preserving lip sync.
Additionally, in assessing the visual quality of the generated talking face videos, this study employs Peak Signal-to-Noise Ratio (PSNR) and Structured Similarity (SSIM) [45] for evaluating pixel-level visual quality. Moreover, we assess feature-level visual quality using the Learned Perceptual Image Patch Similarity (LPIPS) [46] and Fréchet Inception Distance (FID) [47]. Compared to pixel-level measurements, the feature-level evaluations are more aligned with the human perception, as observed in [46, 48].
IV-A4 Comparison Methods
In this work, we compare the proposed method with several representative state-of-the-art (SOTA) subject-generic approaches in audio-driven talking face video generation, including StyleTalk [6], DiffTalk [10], PD-FGC [16], IP-LAP [8], PC-AVS [9], and Wav2Lip [7]. StyleTalk [6] is a representative subject-generic method capable of modeling speaking style. However, the code released for StyleTalk is incomplete and unable to generate results, lacking the phoneme label extraction and training code. Consequently, we made some minor modifications to its code for comparison purposes. The modified version utilizes DeepSpeech [30] audio features instead of phoneme labels and is trained using the same objectives as ours. Additionally, StyleTalk’s image rendering network can only accept a reference facial image to provide subject appearance information, whereas other comparison methods input more conditions, such as the upper-half face at the current time step. For a fair comparison, we replaced StyleTalk’s image renderer with our diffusion rendering model to generate results. We only compare with StyleTalk in terms of lip sync quality. The methods including DiffTalk[10], IP-LAP [8], Wav2Lip[7], and Ours all synthesize talking face videos through inpainting the lower half of the face influenced by the input audio and reference facial image. Therefore, for quantitative comparison, the lower half of the face in the input video is masked and then these methods reconstruct the masked area. The input video before masking serves as the ground truth for metric computation. We train DiffTalk [10] using the released official code until convergence, but it generates temporally unstable lip motion. It relies on additional frame-interpolation techniques to smooth the results, affecting the comparison fairness. Therefore, the frame-interpolation post-processing was not utilized for a fair comparison. PC-AVS [9] utilizes a pose source video, an audio input, and a reference facial image to synthesize a talking face video. For a fair comparison, we replace its pose source video with the ground-truth video. PD-FGC [16] inputs a pose source video, an expression source video, an eye blink source video, audio, and a reference facial image to synthesize a talking face video. For a fair comparison, we replace its pose source, expression source, and eye blink source videos with the ground-truth video.
IV-B Quantitative Evaluation
Method | VoxCeleb | HDTF | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
LipLMD | SyncScore† | PSNR | SSIM | LPIPS | FID | LipLMD | SyncScore† | PSNR | SSIM | LPIPS | FID | ||
Wav2Lip [7] (MM’20) | 0.0180 | 9.49 | 24.34 | 0.82 | 0.168 | 73.2 | 0.0178 | 9.59 | 22.41 | 0.79 | 0.233 | 88.19 | |
PC-AVS [9] (CVPR’21) | 0.1042 | 8.34 | 22.07 | 0.77 | 0.128 | 50.0 | 0.6545 | 8.88 | 19.13 | 0.68 | 0.183 | 55.83 | |
IP-LAP [8] (CVPR’23) | 0.0121 | 3.53 | 24.20 | 0.84 | 0.114 | 40.4 | 0.0118 | 3.56 | 22.37 | 0.83 | 0.119 | 34.35 | |
PD-FGC [16] (CVPR’23) | 0.1529 | 6.41 | 20.26 | 0.70 | 0.165 | 54.8 | 0.1550 | 6.57 | 17.83 | 0.65 | 0.207 | 65.39 | |
DiffTalk [10] (CVPR’23) | 1.3870 | 1.49 | 22.43 | 0.74 | 0.119 | 42.0 | 0.5554 | 2.27 | 22.04 | 0.74 | 0.118 | 30.74 | |
StyleTalk* [6] (AAAI’23) | 0.0119 | 5.37 | - | - | - | - | 0.0125 | 5.56 | - | - | - | - | |
StyleSync (Ours) | 0.0109 | 7.45 | 26.69 | 0.87 | 0.075 | 33.8 | 0.0108 | 6.02 | 24.90 | 0.85 | 0.089 | 26.45 |
We conduct quantitative comparisons with the state-of-the-art methods in terms of lip sync quality and visual quality. Wav2Lip [7], IP-LAP [8], DiffTalk [10], and the proposed approach focuses on generating the lower-half face for lip sync while inheriting the upper-half face from the input video. To be fair, we compute the visual quality metrics solely based on the lower half of the generated face. We report the quantitative comparison results in the Table I.
IV-B1 Lip Sync Quality
As detailed by the LipLMD metric, the proposed method outperforms all of the comparison baselines on both VoxCeleb and HDTF datasets. This demonstrates its generation of more accurate lip sync that better preserves the speaking styles of the individuals. Specifically, Wav2Lip, PC-AVS, DiffTalk, IP-LAP, and PD-FGC lag significantly behind our proposed method in LipLMD due to their incapability to model the speaking styles of individuals. StyleTalk incorporates the use of style reference videos to model speaking styles. However, it still lags behind our StyleSync because of its inaccuracies in style aggregation. Specifically, StyleSync’s LipLMD obtains lower scores than StyleTalk by 8.4% on VoxCeleb and 13.6% on HDTF.
In SyncScore, Wav2Lip and PC-AVS achieve higher scores than the proposed method. This is because Wav2Lip utilizes SyncNet as the lip sync expert to guide the training of its generator, and PC-AVS adopts an audio-visual contrastive learning mechanism similar to SyncNet’s. However, this only indicates that their generated lip shapes are closer to the lip shapes matching the general speaking style rather than the individuals’ speaking styles, since the SyncNet for computing SyncScore overlooks modeling the speaking styles of individuals.
IV-B2 Visual Quality
As shown in the Table I, the proposed StyleSync consistently outperforms other algorithms in terms of visual quality metrics, including PSNR, SSIM, LPIPS, and FID. Particularly noteworthy are the improvements observed in the feature-level metrics LPIPS and FID, where our method demonstrates significant enhancements over other approaches. This suggests that our results align more closely with human perception, demonstrating the efficacy of our conditional latent diffusion model in rendering high-fidelity and realistic faces. As well, DiffTalk devises a latent diffusion framework to generate high-fidelity talking face videos but still lags behind ours. This may be because it simply concatenates the misaligned reference image latent as condition while our method integrates the multi-scale reference image features through spacial cross-attention layers. Furthermore, IP-LAP employs the predicted optical flow to align the reference image and features but still achieves inferior performance than our algorithm in terms of visual quality. This may be due to the training instability and mode collapse of its GAN-based framework.
IV-C Qualitative Evaluation

For a comprehensive qualitative evaluation, we visualize some representative comparison results conducted on HDTF [14] and VoxCeleb [40] datasets in Figure 3 and give some analyses in the following subsections. To facilitate comparisons, we present the cropped faces from the generated results of all methods including the ground-truth face. Moreover, we conduct a user study from volunteers and provide a supplementary video for further evaluation.
IV-C1 Lip Sync Quality
As shown in Figure 3, compared to other algorithms, the proposed method can produce lip shapes that are closer to the ground truth. This indicates our superiority in achieving accurate lip sync conforming to the speaking styles of individuals. StyleTalk sometimes generates inaccurate lip shapes due to its inaccuracy of the style aggregation scheme. Moreover, other approaches, such as DiffTalk, PD-FGC, IP-LAP, PC-AVS, and Wav2Lip, produce sub-optimal lip shapes because they overlook modeling the speaking styles of individuals.
IV-C2 Visual Quality
From Figure 3, it can be seen that the proposed method can generate more realistic faces with high-fidelity facial details and fewer artifacts. Compared to other methods, our results visually resemble real faces more closely. This verifies the effectiveness of the proposed conditional latent diffusion model in rendering lip motion into realistic faces. During implementation, StyleTalk uses the same rendering model as ours, resulting in visual quality similar to ours. Besides, DiffTalk exhibits more artifacts than our approach and loses some facial details of the subject. Moreover, PD-FGC and PC-AVS yield results that compromise the subject’s identity and display significant flaws. Meanwhile, IP-LAP generates facial details that are blurrier than ours and exhibits artifacts particularly under extreme head poses. Similarly, Wav2Lip produces blurry results with unrealistic mouth details.
Method | Lip Sync | Visual Quality |
---|---|---|
Wav2Lip | 3.15 | 2.60 |
PC-AVS | 2.58 | 2.33 |
IP-LAP | 2.19 | 2.35 |
PD-FGC | 2.74 | 2.32 |
DiffTalk | 0.90 | 1.49 |
StyleTalk | 3.93 | - |
Ours | 4.33 | 4.60 |
IV-C3 User Study
To further validate the effectiveness of our method, we conduct a user study from 12 volunteers following the common practice of previous methods. We randomly generate 6 videos of the HDTF [14] and VoxCeleb [40] datasets for each method. The volunteers are asked to give their scores (0-5) for each generated video of all comparison methods regarding lip sync quality and visual quality. We present the mean opinion scores (MOS) of each method in Table II. Our method receives better scores from participants across two dimensions than other methods, indicating the superiority of our method.
IV-C4 Supplementary Video
We have submitted a supplementary video in the review system. The supplementary video demonstrates the comparison results of all methods in several subjects from VoxCeleb [40] and HDTF [14] datasets. Please download and watch it. If there are any issues downloading the video from the review system, the same file can also be downloaded via this single-blind backup link.
IV-D Style Aggregation Analysis

In our method, we posit that within a style reference video, if a lip shape corresponds to the reference audio that is closer to the input audio, it should contribute more significantly to the style aggregation. To verify this assumption, we visualize the attention weights of the cross-attention layers in the transformer decoder during the lip motion prediction stage. There are several cross-attention layers in the transformer decoder. We average the attention weights of all cross-attention layers to obtain a single attention weight, as shown in Figure 4. To facilitate visualization, we set the length of the style reference video to 32 frames and denote the frame numbers as .
As illustrated in Figure 4, the input audio corresponds to the syllable “æ”. The reference audio of the -th frame in the style reference video corresponds to the syllable “e”, which is closer to the input audio’s syllable than the syllables in other frames. Therefore, the attention weight for the -th frame is higher than other frames’, indicating that the reference lip shape in the -th frame contributes more to the style aggregation.
Variant | LipLMD |
---|---|
Ours w/o Reference Audio | 0.01151 |
Ours w/o Vertices Loss | 0.01145 |
Ours w/o Reference (=0) | 0.01552 |
Ours (=64) | 0.01125 |
Ours (=128) | 0.01084 |
Ours (=256) | 0.01081 |
IV-E Ablation Study

To understand the advantage of the core components proposed in our StyleSync, we conduct extensive experiments of ablation study on the HDTF dataset. Here, we only adopt the LipLMD metric to evaluate the lip sync quality. This is because the SyncScore primarily assesses whether the generated lip sync matches the general speaking style, as previously mentioned. We report all numerical ablation results in Table III and present the qualitative results in Figure 5.
IV-E1 Effect of Audio-Aware Style Aggregation
The main contribution of the proposed method is the introduction of the novel audio-aware style reference scheme, which aggregates speaking style information by exploring the relationship between the input audio and the reference audios. Actually, the effectiveness of this has been verified by the fact that the LipLMD score of our method is significantly lower than that of StyleTalk as shown in Table I.
Furthermore, to verify the role of the reference audio, we devise a model variant where the reference audios from the style reference video are not utilized for style aggregation like previous methods [6, 13]. Specifically, the Transformer encoder for encoding the reference audio is removed from the proposed framework, and the reference lip features encoded from the reference lip shapes act as both keys and values for the cross-attention layers in transformer decoder. We report the result of this variant in the row "Ours w/o Reference Audio", and its LipLMD score increases by 6.5% compared to our full model’s. Besides, as shown in Figure 5, without the reference audio, the lip sync quality gets worse. This verifies the necessity to explore the relationship between reference and input audio for accurate and style-preserving lip sync.
IV-E2 Effect of Vertices Loss
As the mouth-related 3DMM expression parameters may involve some information unrelated to lip shape, we apply the reconstruction loss over the lip vertices converted from the predicted expression parameters. To verify the effectiveness of this for enhancing lip sync, we devise a variant where the vertices loss in Equation 4 is not utilized for training the lip motion prediction network. The result of this variant is reported in the row "Ours w/o Vertices Loss" of Table III. The LipLMD increases by 5.9% without using the vertices loss. As well, from Figure 5, the lip sync accuracy becomes worse without the vertices loss for optimization.
IV-E3 Effect of Style Reference Video
To evaluate the effectiveness of incorporating the style reference video for modeling the speaking style of individuals, we introduce a variant that predicts lip motion solely from input audio without utilizing the style reference video. In this variant, the reference branch of our lip motion prediction network is removed, and we replace the cross-attention layers in the transformer decoder with self-attention layers. The performance of this variant is reported by the row labeled “Ours w/o Reference (=0)” in Table III. It is evident that the LipLMD score significantly increases without modeling the speaking style.
Additionally, to assess the impact of the length of the style reference video, we varied during inference and reported results for and in Table III. Increasing the length of the style reference video leads to improved LipLMD metrics. Besides, as shown in Figure 5, with the length of the style reference video decreasing, the lip sync quality gradually deteriorates.
V CONCLUSION
In this paper, we have introduced a novel method called StyleSync to tackle the problem of audio-driven lip sync. It can effectively respond to the challenges of traditional methods that can only achieve sub-optimal lip sync conforming to the general speaking styles or struggle to preserve the speaking styles of individuals. Our proposed audio-aware style reference scheme, leveraging the relationships between input audio and reference audio from style reference video, effectively achieves style-preserving lip sync. In detail, the components developed by StyleSync include a novel Transformer-based lip motion prediction model that integrates speaking style information from a style reference video to accurately predict lip motions, alongside a novel conditional latent diffusion model that translates these predicted motions into realistic, lip-synced videos. These significantly contribute to producing realistic and high-fidelity talking face videos. Experiments as well as comprehensive ablation analysis have revealed StyleSync’s superiority in handling the tasks of audio-driven lip sync.
References
- [1] P. KR, R. Mukhopadhyay, J. Philip, A. Jha, V. Namboodiri, and C. Jawahar, “Towards automatic face-to-face translation,” in Proceedings of the 27th ACM international conference on multimedia, 2019, pp. 1428–1436.
- [2] T. Xie, L. Liao, C. Bi, B. Tang, X. Yin, J. Yang, M. Wang, J. Yao, Y. Zhang, and Z. Ma, “Towards realistic visual dubbing with heterogeneous sources,” in Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 1739–1747.
- [3] Z. Zhou, Z. Wang, S. Yao, Y. Yan, C. Yang, G. Zhai, J. Yan, and X. Yang, “Dialoguenerf: Towards realistic avatar face-to-face conversation video generation,” arXiv preprint arXiv:2203.07931, 2022.
- [4] J. Thies, M. Elgharib, A. Tewari, C. Theobalt, and M. Nießner, “Neural voice puppetry: Audio-driven facial reenactment,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XVI 16. Springer, 2020, pp. 716–731.
- [5] S. Wang, Y. Ma, Y. Ding, Z. Hu, C. Fan, T. Lv, Z. Deng, and X. Yu, “Styletalk++: A unified framework for controlling the speaking styles of talking heads,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024.
- [6] Y. Ma, S. Wang, Z. Hu, C. Fan, T. Lv, Y. Ding, Z. Deng, and X. Yu, “Styletalk: One-shot talking head generation with controllable speaking styles,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 2, 2023, pp. 1896–1904.
- [7] K. Prajwal, R. Mukhopadhyay, V. P. Namboodiri, and C. Jawahar, “A lip sync expert is all you need for speech to lip generation in the wild,” in Proceedings of the 28th ACM international conference on multimedia, 2020, pp. 484–492.
- [8] W. Zhong, C. Fang, Y. Cai, P. Wei, G. Zhao, L. Lin, and G. Li, “Identity-preserving talking face generation with landmark and appearance priors,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 9729–9738.
- [9] H. Zhou, Y. Sun, W. Wu, C. C. Loy, X. Wang, and Z. Liu, “Pose-controllable talking face generation by implicitly modularized audio-visual representation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 4176–4186.
- [10] S. Shen, W. Zhao, Z. Meng, W. Li, Z. Zhu, J. Zhou, and J. Lu, “Difftalk: Crafting diffusion models for generalized audio-driven portraits animation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 1982–1991.
- [11] H. Fu, Z. Wang, K. Gong, K. Wang, T. Chen, H. Li, H. Zeng, and W. Kang, “Mimic: Speaking style disentanglement for speech-driven 3d facial animation,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, no. 2, 2024, pp. 1770–1777.
- [12] S. Tan, B. Ji, Y. Ding, and Y. Pan, “Say anything with any style,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, no. 5, 2024, pp. 5088–5096.
- [13] H. Wu, J. Jia, H. Wang, Y. Dou, C. Duan, and Q. Deng, “Imitating arbitrary talking style for realistic audio-driven talking face synthesis,” in Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 1478–1486.
- [14] Z. Zhang, L. Li, Y. Ding, and C. Fan, “Flow-guided one-shot talking face generation with a high-resolution audio-visual dataset,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 3661–3670.
- [15] L. Chen, Z. Wu, R. Li, W. Bao, J. Ling, X. Tan, and S. Zhao, “Vast: Vivify your talking avatar via zero-shot expressive facial style transfer,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 2977–2987.
- [16] D. Wang, Y. Deng, Z. Yin, H.-Y. Shum, and B. Wang, “Progressive disentangled representation learning for fine-grained controllable talking head synthesis,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 17 979–17 989.
- [17] Y. Ma, S. Wang, Y. Ding, B. Ma, T. Lv, C. Fan, Z. Hu, Z. Deng, and X. Yu, “Talkclip: Talking head generation with text-guided expressive speaking styles,” arXiv preprint arXiv:2304.00334, 2023.
- [18] D. Li, K. Zhao, W. Wang, B. Peng, Y. Zhang, J. Dong, and T. Tan, “Ae-nerf: Audio enhanced neural radiance field for few shot talking head synthesis,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, no. 4, 2024, pp. 3037–3045.
- [19] V. Blanz and T. Vetter, “A morphable model for the synthesis of 3d faces,” in Proceedings of the 26th annual conference on Computer graphics and interactive techniques, 1999, pp. 187–194.
- [20] R. Rombach, A. Blattmann, D. Lorenz, P. Esser, and B. Ommer, “High-resolution image synthesis with latent diffusion models,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2022, pp. 10 684–10 695.
- [21] S. Suwajanakorn, S. M. Seitz, and I. Kemelmacher-Shlizerman, “Synthesizing obama: learning lip sync from audio,” ACM Transactions on Graphics (ToG), vol. 36, no. 4, pp. 1–13, 2017.
- [22] Y. Guo, K. Chen, S. Liang, Y.-J. Liu, H. Bao, and J. Zhang, “Ad-nerf: Audio driven neural radiance fields for talking head synthesis,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 5784–5794.
- [23] R. Huang, P. Lai, Y. Qin, and G. Li, “Parametric implicit face representation for audio-driven facial reenactment,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 12 759–12 768.
- [24] C. Du, Q. Chen, T. He, X. Tan, X. Chen, K. Yu, S. Zhao, and J. Bian, “Dae-talker: High fidelity speech-driven talking face generation with diffusion autoencoder,” in Proceedings of the 31st ACM International Conference on Multimedia, 2023, pp. 4281–4289.
- [25] Z. Ye, Z. Jiang, Y. Ren, J. Liu, J. He, and Z. Zhao, “Geneface: Generalized and high-fidelity audio-driven 3d talking face synthesis,” in The Eleventh International Conference on Learning Representations, 2022.
- [26] B. Mildenhall, P. P. Srinivasan, M. Tancik, J. T. Barron, R. Ramamoorthi, and R. Ng, “Nerf: Representing scenes as neural radiance fields for view synthesis,” Communications of the ACM, vol. 65, no. 1, pp. 99–106, 2021.
- [27] J. Ho, A. Jain, and P. Abbeel, “Denoising diffusion probabilistic models,” Advances in neural information processing systems, vol. 33, pp. 6840–6851, 2020.
- [28] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial networks,” Communications of the ACM, vol. 63, no. 11, pp. 139–144, 2020.
- [29] J. Song, C. Meng, and S. Ermon, “Denoising diffusion implicit models,” in International Conference on Learning Representations, 2020.
- [30] A. Hannun, C. Case, J. Casper, B. Catanzaro, G. Diamos, E. Elsen, R. Prenger, S. Satheesh, S. Sengupta, A. Coates et al., “Deep speech: Scaling up end-to-end speech recognition,” arXiv preprint arXiv:1412.5567, 2014.
- [31] Y. Deng, J. Yang, S. Xu, D. Chen, Y. Jia, and X. Tong, “Accurate 3d face reconstruction with weakly-supervised learning: From single image to image set,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, 2019, pp. 0–0.
- [32] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
- [33] Y. Ma, S. Zhang, J. Wang, X. Wang, Y. Zhang, and Z. Deng, “Dreamtalk: When expressive talking head generation meets diffusion probabilistic models,” arXiv preprint arXiv:2312.09767, 2023.
- [34] S. Wang, L. Li, Y. Ding, and X. Yu, “One-shot talking face generation from single-speaker audio-visual correlation learning,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, no. 3, 2022, pp. 2531–2539.
- [35] A. Bulat and G. Tzimiropoulos, “How far are we from solving the 2d & 3d face alignment problem? (and a dataset of 230,000 3d facial landmarks),” in International Conference on Computer Vision, 2017.
- [36] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. Springer, 2015, pp. 234–241.
- [37] T. Karras, S. Laine, M. Aittala, J. Hellsten, J. Lehtinen, and T. Aila, “Analyzing and improving the image quality of stylegan,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 8110–8119.
- [38] L. Hu, X. Gao, P. Zhang, K. Sun, B. Zhang, and L. Bo, “Animate anyone: Consistent and controllable image-to-video synthesis for character animation,” arXiv preprint arXiv:2311.17117, 2023.
- [39] A. K. Bhunia, S. Khan, H. Cholakkal, R. M. Anwer, J. Laaksonen, M. Shah, and F. S. Khan, “Person image synthesis via denoising diffusion model,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 5968–5976.
- [40] A. Nagrani, J. S. Chung, and A. Zisserman, “Voxceleb: a large-scale speaker identification dataset,” in INTERSPEECH, 2017.
- [41] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
- [42] R. Rombach, A. Blattmann, D. Lorenz, P. Esser, and B. Ommer, “latent-diffusion,” https://github.com/CompVis/latent-diffusion, 2022.
- [43] L. Chen, Z. Li, R. K. Maddox, Z. Duan, and C. Xu, “Lip movements generation at a glance,” in Proceedings of the European conference on computer vision (ECCV), 2018, pp. 520–535.
- [44] J. S. Chung and A. Zisserman, “Out of time: automated lip sync in the wild,” in Asian conference on computer vision. Springer, 2016, pp. 251–263.
- [45] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE transactions on image processing, vol. 13, no. 4, pp. 600–612, 2004.
- [46] R. Zhang, P. Isola, A. A. Efros, E. Shechtman, and O. Wang, “The unreasonable effectiveness of deep features as a perceptual metric,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 586–595.
- [47] M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter, “Gans trained by a two time-scale update rule converge to a local nash equilibrium,” Advances in neural information processing systems, vol. 30, 2017.
- [48] R. Zhen, W. Song, Q. He, J. Cao, L. Shi, and J. Luo, “Human-computer interaction system: A survey of talking-head generation,” Electronics, vol. 12, no. 1, p. 218, 2023.